Search

Andrea Basso Phones & Addresses

  • 3072 SW Blout Ct, Port St Lucie, FL 34953 (718) 864-6906
  • Port Saint Lucie, FL
  • 140 Foch Ave, Staten Island, NY 10305 (718) 816-7279
  • 35 Milbank Rd, Staten Island, NY 10306
  • 861 Avenue Z, Brooklyn, NY 11235 (718) 256-7134
  • Port St Lucie, FL
  • 110 Beverly Ave, Staten Island, NY 10301 (917) 548-4846

Specialities

Civil Litigation

Professional Records

Lawyers & Attorneys

Andrea Basso Photo 1

Andrea Basso - Lawyer

View page
Specialties:
Civil Litigation
ISLN:
913543311
Admitted:
1996
University:
Cornell University, B.A., 1992; Vermont Law School, 1996
Law School:
Vermont Law School, J.D., 1996

Publications

Wikipedia References

Andrea Basso Photo 3

Andrea Basso

Us Patents

Flexible Interchange Of Coded Multimedia Facilitating Access And Streaming

View page
US Patent:
6751623, Jun 15, 2004
Filed:
Jan 26, 1999
Appl. No.:
09/236548
Inventors:
Andrea Basso - Ocean NJ
Alexandros Eleftheriadis - New York NY
Hari Kalva - New York NY
Atul Puri - Riverdale NY
Robert Lewis Schmidt - Howell NJ
Assignee:
ATT Corp. - New York NY
The Trustees of Columbia University - New York NY
International Classification:
G06F 1700
US Classification:
707101, 7071041, 717107, 717108, 382166, 382171, 382233, 37524024, 348 1412, 348 1413
Abstract:
A fundamental limitation in the exchange of audiovisual information today is that its representation is extremely low level. It is composed of coded video or audio samples (often as blocks) arranged in a commercial format. In contrast, the new generation multimedia requires flexible formats to allow a quick adaptation to requirements in terms of access, bandwidth scalability, streaming as well as general data reorganization. The Flexible-Integrated Intermedia Format (Flexible-IIF or F-IIF) is an advanced extension to the Integrated Intermedia Format (IIF). The Flexible-Integrated Intermedia Format (Flexible-IIF) datastructures, file formats systems and methods provide a framework that integrates advanced concepts, such as objects based audio-visual representation, meta-data and object oriented programming, to achieve a flexible and generic representation of the audiovisual information and the associated methods to operate on the audiovisual information.

Method And System For Aligning Natural And Synthetic Video To Speech Synthesis

View page
US Patent:
6862569, Mar 1, 2005
Filed:
Jan 23, 2003
Appl. No.:
10/350225
Inventors:
Andrea Basso - Ocean NJ, US
Mark Charles Beutnagel - Mendham NJ, US
Joern Ostermann - Morganville NJ, US
Assignee:
AT&T Corp. - New York NY
International Classification:
G10L013/00
G06T013/00
US Classification:
704260, 345473, 345706
Abstract:
According to MPEG-4's TTS architecture, facial animation can be driven by two streams simultaneously—text, and Facial Animation Parameters. In this architecture, text input is sent to a Text-To-Speech converter at a decoder that drives the mouth shapes of the face. Facial Animation Parameters are sent from an encoder to the face over the communication channel. The present invention includes codes (known as bookmarks) in the text string transmitted to the Text-to-Speech converter, which bookmarks are placed between words as well as inside them. According to the present invention, the bookmarks carry-an encoder time stamp. Due to the nature of text-to-speech conversion, the encoder time stamp does not relate to real-world time, and should be interpreted as a counter. In addition, the Facial Animation Parameter stream carries the same encoder time stamp found in the bookmark of the text. The system of the present invention reads the bookmark and provides the encoder time stamp as well as a real-time time stamp to the facial animation system.

Scalable Video Encoder/Decoder With Drift Control

View page
US Patent:
6961383, Nov 1, 2005
Filed:
Nov 21, 2001
Appl. No.:
09/683122
Inventors:
Amy Ruth Reibman - Chatham NJ, US
Leon Bottou - Highlands NJ, US
Andrea Basso - Ocean NJ, US
Assignee:
AT&T Corp. - New York NY
International Classification:
H04N007/12
US Classification:
37524025, 37524002, 37524026
Abstract:
Scalable video coders have traditionally avoided using enhancement layer information to predict the base layer, so as to avoid so-called “drift”. As a result, they are less efficient than a one-layer coder. The present invention is directed to a scalable video coder that allows drift, by predicting the base layer from the enhancement layer information. Through careful management of the amount of drift introduced, the overall compression efficiency can be improved while only slighly degrading resilience for lower bit-rates.

Digitally-Generated Lighting For Video Conferencing Applications

View page
US Patent:
6980697, Dec 27, 2005
Filed:
Jan 25, 2002
Appl. No.:
10/057063
Inventors:
Andrea Basso - Ocean NJ, US
Eric Cosatto - Highlands NJ, US
David Crawford Gibbon - Lincroft NJ, US
Hans Peter Graf - Lincroft NJ, US
Shan Liu - Los Angeles CA, US
Assignee:
AT&T Corp. - New York NY
International Classification:
G06K009/40
US Classification:
382274, 382103, 382190, 382291, 348 1401
Abstract:
A method of improving the lighting conditions of a real scene or video sequence. Digitally generated light is added to a scene for video conferencing over telecommunication networks. A virtual illumination equation takes into account light attenuation, lambertian and specular reflection. An image of an object is captured, a virtual light source illuminates the object within the image. In addition, the object can be the head of the user. The position of the head of the user is dynamically tracked so that an three-dimensional model is generated which is representative of the head of the user. Synthetic light is applied to a position on the model to form an illuminated model.

Method And System For Aligning Natural And Synthetic Video To Speech Synthesis

View page
US Patent:
7110950, Sep 19, 2006
Filed:
Jan 7, 2005
Appl. No.:
11/030781
Inventors:
Andrea Basso - Ocean NJ, US
Mark Charles Beutnagel - Mendham NJ, US
Joern Ostermann - Morganville NJ, US
Assignee:
AT&T Corp. - New York NY
International Classification:
G10L 13/00
G06T 13/00
US Classification:
704260, 345473, 715706
Abstract:
According to MPEG-4's TTS architecture, facial animation can be driven by two streams simultaneously—text, and Facial Animation Parameters. In this architecture, text input is sent to a Text-To-Speech converter at a decoder that drives the mouth shapes of the face. Facial Animation Parameters are sent from an encoder to the face over the communication channel. The present invention includes codes (known as bookmarks) in the text string transmitted to the Text-to-Speech converter, which bookmarks are placed between words as well as inside them. According to the present invention, the bookmarks carry an encoder time stamp. Due to the nature of text-to-speech conversion, the encoder time stamp does not relate to real-world time, and should be interpreted as a counter. In addition, the Facial Animation Parameter stream carries the same encoder time stamp found in the bookmark of the text. The system of the present invention reads the bookmark and provides the encoder time stamp as well as a real-time time stamp to the facial animation system.

Digitally-Generated Lighting For Video Conferencing Applications

View page
US Patent:
7231099, Jun 12, 2007
Filed:
Aug 31, 2005
Appl. No.:
11/216997
Inventors:
Andrea Basso - Ocean NJ, US
Eric Cosatto - Highlands NJ, US
David Crawford Gibbon - Lincroft NJ, US
Hans Peter Graf - Lincroft NJ, US
Shan Liu - Los Angeles CA, US
Assignee:
AT&T - New York NY
International Classification:
G06K 9/40
H04N 7/14
US Classification:
382274, 382103, 382190, 382291, 348 1401
Abstract:
A method of improving the lighting conditions of a real scene or video sequence. Digitally generated light is added to a scene for video conferencing over telecommunication networks. A virtual illumination equation takes into account light attenuation, lambertian and specular reflection. An image of an object is captured, a virtual light source illuminates the object within the image. In addition, the object can be the head of the user. The position of the head of the user is dynamically tracked so that an three-dimensional model is generated which is representative of the head of the user. Synthetic light is applied to a position on the model to form an illuminated model.

Method And System For Aligning Natural And Synthetic Video To Speech Synthesis

View page
US Patent:
7366670, Apr 29, 2008
Filed:
Aug 11, 2006
Appl. No.:
11/464018
Inventors:
Andrea Basso - Ocean NJ, US
Mark Charles Beutnagel - Mendham NJ, US
Joern Ostermann - Morganville NJ, US
Assignee:
AT&T Corp. - New York NY
International Classification:
G10L 13/00
G06T 13/00
US Classification:
704260, 704276, 345473
Abstract:
Facial animation in MPEG-4 can be driven by a text stream and a Facial Animation Parameters (FAP) stream. Text input is sent to a TTS converter that drives the mouth shapes of the face. FAPs are sent from an encoder to the face over the communication channel. Disclosed are codes bookmarks in the text string transmitted to the TTS converter. Bookmarks are placed between and inside words and carry an encoder time stamp. The encoder time stamp does not relate to real-world time. The FAP stream carries the same encoder time stamp found in the bookmark of the text. The system reads the bookmark and provides the encoder time stamp as well as a real-time time stamp to the facial animation system. The facial animation system associates the correct facial animation parameter with the real-time time stamp using the encoder time stamp of the bookmark as a reference.

System And Method Of Organizing Data To Facilitate Access And Streaming

View page
US Patent:
7428547, Sep 23, 2008
Filed:
Feb 24, 2004
Appl. No.:
10/785905
Inventors:
Andrea Basso - North Long Branch NJ, US
Alexandros Eleftheriadis - New York NY, US
Hari Kalva - New York NY, US
Atul Puri - Cupertino CA, US
Robert Lewis Schmidt - Howell NJ, US
Assignee:
AT&T Corp. - New York NY
The Trustees of Columbia University - New York NY
International Classification:
G06F 7/00
US Classification:
707101, 707 2, 707100, 707102
Abstract:
File formats systems and methods are disclosed that provide a framework that integrates concepts, such as objects based audio-visual representation, meta-data and object oriented programming, to achieve a flexible and generic representation of the audiovisual information and the associated methods to operate on the audiovisual information. A system and method are disclosed for storing data processed from presentation data. The data is stored according to a method comprising coding input presentation data by identifying objects from within the presentation data, coding each object individually and organizing the coded data into access layer data units. The access layer data units are stored throughout a plurality of segments, each segment comprising a segment table in a header portion thereof and those access layer data units that are members of the respective segment, there being one entry in the segment table for each access layer data unit therein. A plurality of extended segments are also stored, each of the extended segments further comprising one or more of the access layer data units that include protocol specific data, the extended segments each represented by a extended segment header. The data of an accessible object is also stored, including an accessible object header and identifiers of the plurality of extended segments, each of the extended segments being a member of the same object.
Andrea L Basso from Port Saint Lucie, FL, age ~60 Get Report