US20140055401A1 - Method and apparatus for controlling contents in realtime fashion - Google Patents

Method and apparatus for controlling contents in realtime fashion Download PDF

Info

Publication number
US20140055401A1
US20140055401A1 US14/069,384 US201314069384A US2014055401A1 US 20140055401 A1 US20140055401 A1 US 20140055401A1 US 201314069384 A US201314069384 A US 201314069384A US 2014055401 A1 US2014055401 A1 US 2014055401A1
Authority
US
United States
Prior art keywords
screen
touch panel
position information
touch
network communication
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US14/069,384
Inventor
Zhou Ye
Pei-Chuan Liu
San-Yuan Huang
Yun-Fei Wei
Ying-Ko Lu
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Bluespace Corp
Original Assignee
Bluespace Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Bluespace Corp filed Critical Bluespace Corp
Priority to US14/069,384 priority Critical patent/US20140055401A1/en
Assigned to BLUESPACE CORPORATION reassignment BLUESPACE CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: HUANG, SAN-YUAN, LIU, Pei-chuan, LU, YING-KO, WEI, Yun-fei, YE, ZHOU
Publication of US20140055401A1 publication Critical patent/US20140055401A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/041Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/14Digital output to display device ; Cooperation and interconnection of the display device with other functional units
    • G06F3/1454Digital output to display device ; Cooperation and interconnection of the display device with other functional units involving copying of the display data of a local workstation or window to a remote workstation or window so that an actual copy of the data is displayed simultaneously on two or more displays, e.g. teledisplay
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2330/00Aspects of power supply; Aspects of display protection and defect management
    • G09G2330/02Details of power systems and of start or stop of display operation
    • G09G2330/021Power management, e.g. power saving
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2340/00Aspects of display data processing
    • G09G2340/02Handling of images in compressed format, e.g. JPEG, MPEG
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2350/00Solving problems of bandwidth in display systems
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2354/00Aspects of interface with display user
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2370/00Aspects of data communication
    • G09G2370/16Use of wireless transmission of display information

Definitions

  • the present invention relates to a method and an apparatus for controlling contents in realtime fashion, and more particularly, to a method and an apparatus for controlling multimedia contents in realtime fashion by means of using wireless touch, motion sensing and battery charging.
  • each one of such devices may be operated under a specific platform or OS that may be unique from each other, which may be good in providing varieties to the market in general but may too raise a preference issue to the user depending on his/her own demands in terms of such as function, design or performance of different operating systems and devices.
  • the inventor realizes an user may advantageously be provided with an universal apparatus capable of communicating with the any example of new breed of electronic devices running any certain type of operating systems and may be advantageously free from the preference selection issue while still being able to utilize or interact with the any one such new breed of electronic devices running different type of operating systems.
  • portable devices are integrated with more functions to be easily accessed by and interacted with users for more dedicated purposes or services including such as gaming, web browsing, teleconferencing and navigation.
  • the inventor realizes that in order to provide an electronic apparatus capable of communicating with external devices of various of different operating systems offering more dedicated services or applications, the most accessible functions or main services of the communicated external devices shall be preferably preserved, controlled and/or interacted by users while preserving a relatively less hardware structures to provide a highly portable and interactive devices to users.
  • an electronic apparatus capable of communicating with external device(s) of various of different types of operating systems such that applications and services of the connected/communicated external device of different types of operating systems in relation to such as interactive gaming, teleconferencing, web browsing, media playing, navigation, image displaying or taking may all be preserved or advantageously made available, and preferably with enhanced data transmission, while realizing or having a relatively “light” hardware structures with great portability and interactive feasibility to users.
  • the present invention provides a method for performing an actual operation on a second device having a second screen from a first device having a first screen in real time, in which the first device is coupled to the second device via a network communication, the first screen is configured to display a mirrored content of the second screen, and the first screen includes touch panel capability.
  • the method for performing the actual operation on the second device having the second screen from the first device having the first screen in real time includes the following steps: detecting a first touch signal on the first screen of the first device; converting the first touch signal into a first touch data associated with a first position information with respect to a first coordinate system of the first screen; transmitting the first touch data to the second device via the network communication, so that the actual operation is performed on the second device.
  • the actual operation performed on the second device can be overridden if a detecting event is detected on the second screen of the second device.
  • the transmit (TX) side is also referred to as a second device
  • the receive (RX) side is also referred to a first device.
  • optional steps of the embodiment include of the following: detecting a motion event on the first device and generating a sensor data, and sending the sensor data to the second device via the network communication for enabling a corresponding operation on the second device, in which the corresponding operation includes changing orientation of the content displayed on the second screen.
  • one of a plurality of sensing ranges can be selected for detecting the motion event, and a threshold value can be provided for detecting the motion event on the first device.
  • the method further includes charging capability to charge the first device with the second device or to charge the second device with the first device by detecting a charging connection between the first device and the second device, retrieving remaining battery amounts from the first device and the second device respectively, and providing an option to charge the first device with the second device, or to charge the second device with the first device.
  • a method for performing an actual operation on a second device having a second screen from a first device having a first screen in real time further includes calculating a second position information with respect to a second coordinate system of the second screen based on the first position information.
  • the method for performing the actual operation on the second device having the second screen from the first device having the first screen in real time of the embodiment further includes additional steps which include: receiving compressed multimedia content compressed with a compression ratio from the second device via the network communication, and decompressing and displaying the multimedia content on the first screen of the first device.
  • the compression ratio is determined based on a bandwidth of the network communication.
  • the decompressed multimedia content can be scaled based on a resolution of the first screen.
  • the present invention provides a method for performing an actual operation on a second device having a second touch panel from a first device having a first touch panel in real time, and the first device is coupled to the second device via a network communication.
  • the method for performing the actual operation on the second device having the second touch panel from the first device having the first touch panel in real time includes steps as follow: receiving a first touch data associated with a first position information with respect to a first coordinate system of the first touch panel; calculating a second position information with respect to a second coordinate system of the second touch panel based on the first position information; and performing the actual operation on the second device based on the second position information.
  • the transmit (TX) side is also referred to as a second device
  • the receive (RX) side is also referred to a first device.
  • the performing of the actual operation is overridden if a detecting event is detected on the second touch panel of the second device.
  • the method of this embodiment may further include steps for compressing and decompressing of multimedia content from the second device to the first device as follow: compressing a multimedia content displayed on the second touch panel, and sending the compressed multimedia content from the second device to the first device for decompressing and displaying on the first touch panel via the network communication.
  • This embodiment may further include steps for performing operations on the second device in response to sensor data indicating a motion event on the first device as follow: receiving a sensor data from the first device, the sensor data indicating a motion event on the first device, and performing a corresponding operation on the second device in response to the sensor data.
  • the corresponding operation may include changing orientation of the multimedia content displayed on the second display.
  • the present invention provides a second device configured to be controlled by a first device with a first touch panel, and the first device is coupled to the second device by a network communication.
  • the second device includes a second touch panel, a compression unit for compressing multimedia content with a compression ratio, a transceiver for transmitting the compressed multimedia content to the first device for decompressing and displaying on the first touch panel and for receiving a touch data associated with a first position information with respect to a first coordinate system of the first touch panel, a coordinate calculating unit for calculating a second position information with respect to a second coordinate system of the second touch panel, and a processor for performing an actual operation on the second device according to the touch data and the second position information.
  • the actual operation can be overridden if the second touch panel detects a detecting event or if the second device receives an incoming call.
  • the compression ratio is determined based on a bandwidth of the network communication.
  • the transmit (TX) side is also referred to as a second device, and the receive (RX) side is also referred to a first device; in addition, the first touch data is referred to the touch data being formed after converting of the at least one touch signal into touch data.
  • FIG. 1A is a schematic diagram showing wireless control from a RX side over a TX side according to an embodiment of the present invention.
  • FIG. 1B and FIG. 1C are schematic diagrams showing touch control operations of from the RX side over the TX side shown in FIG. 1A according to an embodiment of the present invention.
  • FIG. 2 is a schematic diagram showing a compression ratio in relation to an available bandwidth between the RX side and the TX side according to an alteration of the present invention.
  • FIG. 3A and FIG. 3B are schematic diagrams showing an image orientation on the TX side in response to a motion detection on the RX side where FIG. 3B is further provided with a switch element for adjusting motion sensitivity according to another alteration of the present invention.
  • FIG. 3C is a schematic diagram showing motion control operations of the RX side over the TX side corresponding to FIGS. 3A and 3B according to an embodiment of the present invention.
  • FIG. 3D and FIG. 3E are schematic diagrams showing adjustment of motion sensitivity on the RX side implemented by an application program on the TX side according to a further alteration of the present invention.
  • FIG. 3F and FIG. 3G are schematic diagrams showing adjustment of magnitude and/or threshold values of application programs on the TX side after motion data is received according to a further alteration of the present invention.
  • FIG. 4 is a flow diagram showing a process for controlling multimedia contents in realtime fashion according to an embodiment of the present invention.
  • FIG. 5 is a flow diagram showing a process for compressing and decompressing multimedia contents in realtime fashion according to an embodiment of the present invention.
  • FIG. 6 is a flow diagram showing another process for compressing and decompressing multimedia contents in realtime fashion according to an embodiment of the present invention.
  • FIG. 7A and FIG. 7B are schematic diagrams showing battery charging between the TX side and the RX side according to a further alteration of the present invention.
  • FIG. 8 is a flow diagram of a battery charging process according to an embodiment of the present invention.
  • FIG. 9 is a flow diagram of a battery charging process according to another embodiment of the present invention.
  • FIG. 1A is a schematic diagram of a RX side 10 (e.g. pad-like device having a touch-control display) and a TX side 12 (e.g. phone-like device having a computing power & operating system therein) according to an embodiment of the present invention.
  • the RX side 10 is provided for controlling in realtime fashion multimedia contents that are originated from the TX side 12 .
  • the multimedia contents can include audio and video streams and be processed to present at the TX side 12 (such as cellphone, smartphone, PDA, and GPS navigation, or any other portable wireless device) having a phone-like screen 1202 and the TX side 12 is coupled to the RX side 10 via a network communication NC, which is preferably a wireless communication and a two-way communication.
  • NC is preferably a wireless communication and a two-way communication.
  • the multimedia contents may include video and audio data, i.e. the video data can be displayed on the screen 1202 while the audio data can be played via a speaker (not shown) in the TX side 12 .
  • the RX side 10 may be advantageously provided in a CPU-less (without central processing unit) and/or GPU-less (without graphical processing unit) environment for the sake of less power consumption and/or simplified circuit complexity.
  • the TX side 12 may be a computing system or mobile device equipped with CPU and/or GPU computing power, such as a PC, laptop, tablet PC, mobile phone, smart-phone, and may include a processor (not shown) for executing an application program (not shown) of an operation system (OS) 1204 installed in a storage medium (not shown) and a transceiver 1206 coupled to the processor and capable of transmitting and receiving the multimedia contents generated by the application program of OS 1204 executed by the processor.
  • OS operation system
  • the OS 1204 may be any one of existing or known OS, such as the abovementioned Windows®, iOS®, Android®, Linux®, Blackberry®, Symbian®, Palm®, installed on the TX side 12 such that an application program may be performed or run thereon.
  • the network communication NC between the RX side 10 and the TX side 12 may preferably be any one of the network communication standards or profiles of: IEEE 802.11, IEEE 802.16, CDMA2000, WCDMA, TD-SCDMA, LTE, TDLTE, Bluetooth, Zigbee and combination thereof, or wired communication such as ADSL network, Optical/Fiber network or Cable network.
  • the RX side 10 (refer back to FIG. 1A ) includes a pad-like screen 1002 (e.g. touch panel), a coordinate mapping unit 1004 and a transceiver 1006 .
  • the screen 1002 is provided for detecting at least one touch signal, and is preferably larger than the screen 1202 .
  • the coordinate mapping unit 1004 can be controlled and/or controllable by an operating system (not shown).
  • the coordinate mapping unit 1004 is provided for converting the at least one touch signal (from the screen 1002 ) into touch data associated with a first position information that indicates a virtual operation on the first screen 1002 corresponding to an actual operation on the second screen 1202 , wherein the first position information is with respect to a first coordinate system of the first screen.
  • a pad-like screen 1002 e.g. touch panel
  • the coordinate mapping unit 1004 can be controlled and/or controllable by an operating system (not shown).
  • the coordinate mapping unit 1004 is provided for converting the at least one touch signal (from the screen 100
  • the first position information represents a current cursor position (right side of FIG. 1B ) on the pad-like screen 1202 moved by the user's finger (not shown), and thus the mirrored cursor on the phone-like screen 1202 will be moved to the corresponding position (left side of FIG. 1B ) such that the user can do any operation on the phone-like screen by touching the pad-like screen 1202 while without touching the phone-like screen 1202 .
  • the transceiver 1006 is provided for transmitting the touch data (e.g. the first position information indicative of the current position of cursor on screen 1002 ) sent from the coordinate mapping unit 1004 to the TX side 12 via the network communication NC.
  • the TX side 12 further includes a coordinate calculating unit 1208 for calculating a second position information (e.g. new cursor position on phone-like screen) with respect to a second coordinate system of the screen 1202 based on the first position information.
  • a second position information e.g. new cursor position on phone-like screen
  • mapping movement (delta_x′, delta_y′) on the phone-like screen can be obtained as following expression:
  • the first position information may further include the current position of the cursor on the pad-like screen 1002 of the RX side 10 , and the resolution data of the RX side 10 , e.g. 1920 ⁇ 1080, and the second position information can be calculated so as to obtain the mapping position of the cursor on the phone-like screen 1202 , e.g. scaling down to a resolution 480 ⁇ 320.
  • the actual action can be “play a video movie” when the mapping position of the cursor on the screen 1202 is pointed on a “play” icon.
  • the user can “touch” on the pad-like screen 1002 of the RX side 10 to implement operations without access to the phone-like screen 1202 as if what he/she usually does on the phone-like screen 1202 of the TX side 12 .
  • all the tasks requiring extensive computing power/resources are handled by the TX side 12 , and only the display task is duplicated (mirrored) to the RX side 10 from the TX side 12 such that the multimedia contents can be presented in realtime fashion between the pad-like screen 1002 of the RX side and the phone-like screen 1202 of the TX side 12 .
  • FIG. 1B and FIG. 1C are schematic diagrams of touch control operations of the RX side 10 and the TX side 12 shown in FIG. 1A according to an embodiment of the present invention.
  • a user can do the “open” operation to open a file or folder on the larger screen 1002 of the RX side 10 (right side) and quasi-simultaneously the smaller screen 1202 of the TX side 12 (left side) will mirror the same scene/image as that in the RX side 10 , and thus the user can do the “double-click” operation to open some object like file/folder on the larger screen 1002 while utilize computing power from the TX side 12 , i.e.
  • a user can do the “key-in” operation on the larger RX side 10 (right side) to type words and simultaneously the smaller screen 1202 of the TX side 12 (left side) will mirror the same scene/image as that in the RX side 10 , and thus the user can do the “key-in” operation on the larger screen 1002 while utilize computing power from the TX side 12 , i.e. the operating system 1204 , to run the text-editing application.
  • the TX side 12 may further include a compressing unit 1210 , for compressing the multimedia contents with a compression ratio, e.g. compressing RGB data into MJPEG data, which requires less time for compression and thus facilitates operating in realtime fashion, and the transceiver 1206 can be further utilized for transmitting the compressed multimedia contents to the screen 1002 for displaying.
  • a compressing unit 1210 for compressing the multimedia contents with a compression ratio, e.g. compressing RGB data into MJPEG data, which requires less time for compression and thus facilitates operating in realtime fashion
  • the transceiver 1206 can be further utilized for transmitting the compressed multimedia contents to the screen 1002 for displaying.
  • the transceiver 1006 receives the compressed multimedia contents from the transceiver 1206 via the network communication NC, and then compressed multimedia contents are decompressed by a decompressing unit 1010 with the compression ratio, such that the video data of the multimedia contents can be displayed on the screen 1002 while the audio data of the multimedia contents can be played by a speaker 1008 .
  • the multimedia contents can be displayed and controlled in realtime fashion between the screen 1002 of the RX side 10 and the screen 1202 of the TX side 12 .
  • FIG. 2 is a schematic diagram of the RX side 10 and the TX side 12 according to an alteration of the present invention.
  • the transceiver 1206 can further detect a bandwidth of the network communication NC
  • the TX side 12 further includes a dynamic compression unit 1212 , for determining the compression ratio of the multimedia content based on the bandwidth of the network communication NC, such that the compressing unit 1210 can compress the multimedia contents at the compression ratio, which is determined by the bandwidth of the network communication NC.
  • the transceiver 1006 also can detect the bandwidth of the network communication NC, and then the decompressing unit 1010 decompresses the compressed multimedia contents by the compression ratio provided by a bandwidth calculating unit 1012 .
  • the multimedia contents can be compressed with a compression ratio according to a bandwidth of the network communication NC, such that the multimedia contents can be smoothly displayed in realtime fashion between the screen 1002 of the RX side 10 and the screen 1202 of the TX side 12 by facilitating the realtime codec function (e.g. the video compression ratio is determined by the actual available bandwidth).
  • the transceiver 1006 also can directly receive the compression ratio from the dynamic compression unit 1212 (TX side) rather than the bandwidth calculating unit 1012 , and then to pass the compression ratio into the decompressing unit 1010 so that the bandwidth calculating unit 1012 can be removed as well.
  • the present invention discloses another method for controlling in realtime fashion multimedia contents between a RX side having a first touch panel and a TX side having a second touch panel via a network communication.
  • the multimedia contents comprise at least a screen frame data (not shown) where the screen frame data includes the screen resolution data, image data displayed on the screen and audio data.
  • the method comprising the steps of:
  • FIGS. 3A-3B are schematic diagrams of the TX side 12 and the RX side 10 according to another alteration of the present invention.
  • the RX side 10 may further include a motion sensor 1014 (including at least one of an accelerometer, a gyroscope and a magnetic compass) for detecting at least one motion event.
  • the receiver 1006 transmits the sensor data (RX) from the motion sensor 1014 after the motion sensor 1014 has detected the motion event on the RX side, and then the sensor data (RX) is sent to the TX side 12 via the network communication NC.
  • a sensor decision unit may disable sensor data (TX) generated from a motion sensor 1216 on the TX side 12 only if the detecting exists (i.e. the sensor data (RX) from the RX side has being received on the TX side), and then send the sensor data (RX) to a motion process unit 1214 .
  • the motion process unit 1214 can convert the sensor data (RX) into motion data (indicating a translation representative of linear acceleration in three-axis directions, and/or an angular rotation representative of Yaw, Roll and Pitch in three-axis directions) on the TX side 12 , so as to present the multimedia contents on the screen 1202 and subsequently on the screen 1002 in a same particular orientation (e.g. portrait orientation or landscape orientation).
  • the particular orientation is determined by the operating system 1204 according to the motion data (RX) such that the image at the screen 1202 will be presented at the landscape or portrait orientation.
  • RX motion data
  • the multimedia contents can be displayed and controlled in realtime fashion between the screen 1002 of the RX side 10 and the screen 1202 of the TX side 12 through motion operations.
  • FIG. 3C is a schematic diagrams of motion control operations between the RX side 10 and the TX side 12 shown in FIGS. 3A-3B according to one embodiment of the present invention.
  • a user can do the “game play” driving motion, e.g. moving the image (RX) of the larger pad-like screen 1002 from Landscape to Portrait, and thus the mirrored image (TX) on the smaller phone-like screen 1002 (shown on left side of FIG. 3C ) is quasi-simultaneously adjusted from Landscape to Portrait on the TX side 12 performed by the operating system 1204 based on the motion data converted from the sensor data (RX) generated by the motion sensor 1014 (RX).
  • the user can play the game on the pad-like screen 1002 while utilize computing power from the TX side 12 , i.e. the operating system 1204 , to run the game application actually on the phone-like screen 1202 .
  • the motion sensor 1014 may have a plurality of sensing ranges, i.e. different sensing sensitivities, or a particular sensing range for generating the sensor data.
  • one of the multiple sensing ranges can be selected by a control code determined by a switch signal so as to enable a motion chip to generate one motion data based on the selected sensing range determined by the control code.
  • the motion sensor 1014 has a plurality of sensing ranges. As shown in FIG. 3A-3B , the RX side 10 detects whether a switch signal is invoked by a switch element 1016 operative to a user, and the switch element 1016 may issue an adjustment signal to a motion chip (i.e. motion sensor 1014 ) after the switch signal was invoked (e.g. user turns the switch element 1016 on). Then, the motion sensor 1014 selects one of the plurality of sensing ranges of the motion chip based on the adjustment signal, and generates the motion data by the motion chip based on the selected sensing range so as to enable an interaction between the RX side 10 and the TX side 12 .
  • a motion chip i.e. motion sensor 1014
  • the motion chip has a sensing range of ⁇ 1 G at its “off” (original) state and ⁇ 2 G at “on” state.
  • the motion chip may have a new sensing range of ⁇ 2 G, where G is defined as 9.81 m/s 2 for gravitational acceleration.
  • the motion sensor 1014 has a plurality of sensing ranges. As shown in FIG. 3D-3E , the TX side 12 detects whether a switch signal is invoked by an application program 1220 (e.g. game software/program), and an adjustment signal is generated in response to the invoked switch signal. Then, the motion sensor 1014 selects one of the plurality of sensing ranges of the motion chip of the motion sensor 1014 based on the adjustment signal, and generates the motion data by the motion chip based on the selected sensing range so as to enable an interaction between the RX side 10 and the TX side 12 .
  • an application program 1220 e.g. game software/program
  • the game software may provide an input user interface to allow the user to choose one of three options for selecting particular sensing range among the plurality of sensing ranges where Option 1 stands for ⁇ 1 G; Option 2 for ⁇ 2 G; and Option 3 for ⁇ 3 G, respectively.
  • a particular switch signal will be sent out to the RX side after the selected Option 2 is chosen by the user, for instance.
  • the motion sensor 1014 (RX) is requested to change its sensing range to ⁇ 2 G directly by the switch signal, or by an adjusted signal in response to the switch signal.
  • the motion sensor 1014 has a particular sensing range for generating the sensor data.
  • the TX side 12 detects whether a switch signal is invoked by an application program 1222 of the operating system 1204 , and an adjustment signal is sent in response to the invoked switch signal. Then, the TX side 12 may provide a user interface to allow the user to adjust a threshold value of the application program 1222 (e.g. game software/program) based on either the adjustment signal or the invoked switch signal such that the threshold value of the application program 1222 can be increased at a ratio corresponding to the adjustment signal or the invoked switch signal. Subsequently, the motion data may be generated based on sensor data that is sent from the motion sensor 1014 (on the TX side) triggered by the switch signal, so as to enable an interaction between the RX side 10 and the TX side 12 .
  • a threshold value of the application program 1222 e.g. game software/program
  • the motion sensor 1014 has a particular sensing range for generating the sensor data.
  • the TX side 12 detects whether a switch signal is invoked by the application program 1222 of the operating system 1204 , and issues an adjustment signal in response to the invoked switch signal. Then, the TX side 12 may provide a user interface to allow the user to adjust a magnitude of the motion data by means of increasing the magnitude's value at a ratio based on either the adjustment signal or the switch signal, so as to enable an interaction between the RX side 10 and the TX side 12 .
  • the TX side 12 can be configured to adjust the threshold value of the application program 1222 and the magnitude of the motion data, respectively, but the TX side 12 can also simultaneously adjust the threshold value of the application program 1222 and the matmitude of the motion data, while retaining both merits.
  • the spirit of the present invention is to utilize a RX side to control multimedia contents played at transmitting entities of different types of operating systems in realtime fashion, wherein the RX side only transmits touch data and sensor data for a TX side to perform actual operation, and displays the multimedia contents of the TX side.
  • the RX side 10 can be foldable, portable, or any other types to meet each user's need(s).
  • the network communication NC may preferably include a Wi-Fi communication and a Bluetooth (BT) communication, and since the touch data and sensor data have less data volume while the multimedia contents have greater data volume, the touch data and the motion date are preferably transmitted to the TX side 12 via the BT communication, while the multimedia contents are preferably transmitted to the RX side 10 via the Wi-Fi communication.
  • the transceiver 1006 and 1206 of the present invention may be a wireless module adapting other communication standards or profiles including such as IEEE 802.16 (WiMAX), CDMA2000, WCDMA, TD-SCDMA, LTE, TDLTE, Bluetooth-profile and combination thereof or a wired module.
  • the video data of the multimedia contents decompressed and outputted by the decompressing unit 1010 may preferably from a MJPEG format to a RGB format and transmitted or outputted to the screen 1002 in a LVDS format via a display interface converter such as a RGB-to-LVDS converter. It can be understood that other video stream formats such as ITDMI and related interfaces may also be possible.
  • the audio data of the multimedia contents processed and outputted may preferably be transmitted via an audio interface such as an I 2 S interface or standard and further to the speaker 1008 .
  • the multimedia contents transmitted and processed in video streaming may be made “real-time” or without significant delays with respect to time for the sake of low latency.
  • the multimedia contents may be communicated between the RX side 10 and the TX side 12 and processed to output “real-time” video and audio streams to the screen 1002 and/or the speaker 1008 of the RX side 10 of the present invention.
  • the multimedia contents may be processed in video streaming (i.e.
  • encoding and decoding and communicated there between as data packets and the duration or delay of time involved in a “real-time” processing and transmission may be less than 100 ms for the sake of low latency; in one example, it may preferably be less than 10 ms, and in another example, it may approximately be 20 ms.
  • the controlling process 40 includes the following steps:
  • the receiving process 50 includes the following steps:
  • the receiving and transmitting process 60 includes the following steps:
  • FIG. 7A-7B is a schematic diagram of the RX side 10 according to another alteration of the present invention.
  • the TX side 12 further comprises a charging application 1224 , charging detector 1226 and battery cells 1228
  • the RX side 10 further comprises a charging controller 1018 and battery cells 1020 .
  • system default charging operations of the TX side 12 and the RX side 10 shown in FIG. 7A-7B can be summarized into charging process 80 as shown in FIG. 8 .
  • the charging process 80 includes the following steps:
  • the charging process 90 includes the following steps:
  • the event of a charging connection can be detected either by a charging detector 1226 on the TX side 12 or by a charging controller 1018 on the RX side 10 .
  • the charging controller 1018 can be implemented by a power management chip.
  • the charging detector 1226 can receive the remaining battery capacity of the battery cells 1228 on the TX side 12 ; so as to send the battery capacity information to the user interface by means of the charging application 1224 managed by OS 1204 , and thus users can view the icon that indicates the actual measurement of the TX's battery capacity.
  • the charging controller 1218 on the RX side 10 can receive the actual measurement of the RX's battery capacity, and sends the measurement to the charging detector 1226 on the TX side 12 , so as to facilitate the charging detector 1226 to determine what condition can be meet to implement the appropriate charging operation.
  • a power adaptor is used to charge both the RX side 10 and the TX side 12 ; when the remaining battery amount of the RX side 10 is higher than its predetermined battery capacity percentage, and simultaneously the remaining battery amount of the RX side 10 is higher than that of the TX side 12 (i.e. Rr>Tr and Rr>Rt), the RX side 10 is used to charge the TX side 12 ; when the remaining battery amount of the TX side 12 is higher than its predetermined battery capacity percentage, and simultaneously the remaining battery amount of the TX side 12 is higher than that of the RX side 10 (i.e.
  • the TX side 12 is used to charge the RX side 10 . It is noted that the RX side may be compulsorily switched to the power saving mode by the override mechanism when the user receives an incoming call from the TX side, that is, the user currently uses the TX side rather the RX side to either answer or dial the call. Therefore, it makes sense for the user to temporarily keep the RX side idle while he or she is busy for phone call.
  • the present invention utilizes a RX side to control multimedia contents played at transmitting entities of different types of operating systems in realtime fashion, wherein the RX side only transmits touch data and sensor data for a TX side to perform actual operation, and displays the multimedia contents of the TX side. Furthermore, the present invention can compress multimedia contents a bandwidth of network communication, such that the multimedia contents can be smoothly displayed in realtime fashion between the screen 1002 of the RX side 10 and the screen 1202 of the TX side 12 .
  • the TX side 12 can also be configured as a server having encoding hardware and the RX side 10 as a client having decoding hardware adapted for cloud computing environment running video streaming for the purpose of low latency requirement.

Abstract

Method for performing actual operation on a second device having a second screen or a touch panel from a first device having a first screen or a touch panel in real time are provided. First screen is configured to display mirrored content of second screen. Method includes detecting a first touch signal on first screen of first device, converting first touch signal into a first touch data associated with a first position information, and transmitting first touch data to second device so actual operation is performed on second device. For first and second devices having touch panels, method includes receiving a first touch data associated with a first position information of first touch panel, calculating a second position information of second touch panel based on first position information, and performing actual operation on second device based on second position information. Second device is coupled to first device by network communication.

Description

    CROSS REFERENCE TO RELATED APPLICATIONS
  • This application is a continuation application of U.S. non-provisional application Ser. No. 13/184584, filed on Jul. 18, 2011 and now pending, which in turn claims the benefit of U.S. provisional application Ser. No. 61/410,371, filed on Nov. 05, 2010, now abandoned, the contents of which are incorporated herein by reference and made a part of this disclosure.
  • BACKGROUND OF THE INVENTION
  • 1. Field of the Invention
  • The present invention relates to a method and an apparatus for controlling contents in realtime fashion, and more particularly, to a method and an apparatus for controlling multimedia contents in realtime fashion by means of using wireless touch, motion sensing and battery charging.
  • 2. Description of the Prior Art
  • As the technology related to portable electronic device advances and the need for portability increases in relation to the current market trend, traditional personal computers are transforming to lighter and more dedicated devices provided for or targeted at different purposes and user groups including such as gaming, telecommunication, web browsing and media streaming. Such shift of market and technology trends related to advanced needs of portability and interactions demanded by users of such electronic devices are in fact facilitating not only the improvement of traditional personal computers, laptops or smart-phones to be more powerful and/or integrated with more features and functions than ever but also a new category or breed of portable electronic devices with more dedicated functions for certain purposes and greater portability at the same time.
  • In view of the merits of greater portability as well as interaction provided by such newly evolved breed of electronic devices previously mentioned, the inventor realizes that electronic designers and manufacturers are introducing a “lighter” version of portable personal computers or laptop to the market while telecommunication device designers and manufacturers are, in general, offering a “heavier” version of mobile phones such as smart-phones such that the new growing market and user demand of high portability as well as interactive functions may be fulfilled. Some known examples of such new breed of devices may include: iPad by Apple®, Playbook by RIM®, Cius by CISCO®, Galaxy Tab by Samsung®, Libretto by Toshiba®; and examples of known smart phones may be: iPhone by Apple®, Blackberry Bold by RIM®, Galaxy S by Samsung®, HD7 by HTC™. In general, these portable devices share some common building blocks of traditional personal computers but with greater portability in terms of their size, weight as well as with greater interactive and integrated functions with more dedicated applications and telecommunication capabilities including via such as known 3G, 4G or Wi-Fi™.
  • As such new breed of portable electronic devices are being introduced to the market, the inventor also realizes that each one of such devices may be operated under a specific platform or OS that may be unique from each other, which may be good in providing varieties to the market in general but may too raise a preference issue to the user depending on his/her own demands in terms of such as function, design or performance of different operating systems and devices. Instead of choosing one particular type of portable electronic device miming a certain type of operating systems including such as Windows® by Microsoft®, iOS® by Apple®, Android® or Linux® by Google®, Blackberry® OS by RIM®, the inventor realizes an user may advantageously be provided with an universal apparatus capable of communicating with the any example of new breed of electronic devices running any certain type of operating systems and may be advantageously free from the preference selection issue while still being able to utilize or interact with the any one such new breed of electronic devices running different type of operating systems.
  • However, there are at least two major concerns to be considered in providing an electronic apparatus capable of advantageously communicating and interacting with the newly introduced breed of portable electronic devices, one is related to the various different platforms or operating systems being adapted by different electronic devices and the other is to preserve the merits of great portability and being light in terms of operation and hardware structures or components of portable electronic devices demanded by the market trend. Therefore, there is a need to provide an electronic apparatus capable of communicating with external device(s) of various types of operating systems while preserving a relatively “light” hardware structures to provide a highly portable and interactive devices to users.
  • Furthermore, as previously mentioned that portable devices are integrated with more functions to be easily accessed by and interacted with users for more dedicated purposes or services including such as gaming, web browsing, teleconferencing and navigation. The inventor realizes that in order to provide an electronic apparatus capable of communicating with external devices of various of different operating systems offering more dedicated services or applications, the most accessible functions or main services of the communicated external devices shall be preferably preserved, controlled and/or interacted by users while preserving a relatively less hardware structures to provide a highly portable and interactive devices to users.
  • In view of the foregoing, it is desirable to provide an electronic apparatus capable of communicating with external device(s) of various of different types of operating systems such that applications and services of the connected/communicated external device of different types of operating systems in relation to such as interactive gaming, teleconferencing, web browsing, media playing, navigation, image displaying or taking may all be preserved or advantageously made available, and preferably with enhanced data transmission, while realizing or having a relatively “light” hardware structures with great portability and interactive feasibility to users.
  • SUMMARY OF THE INVENTION
  • In one aspect, the present invention provides a method for performing an actual operation on a second device having a second screen from a first device having a first screen in real time, in which the first device is coupled to the second device via a network communication, the first screen is configured to display a mirrored content of the second screen, and the first screen includes touch panel capability.
  • In an embodiment, the method for performing the actual operation on the second device having the second screen from the first device having the first screen in real time includes the following steps: detecting a first touch signal on the first screen of the first device; converting the first touch signal into a first touch data associated with a first position information with respect to a first coordinate system of the first screen; transmitting the first touch data to the second device via the network communication, so that the actual operation is performed on the second device. In this embodiment, the actual operation performed on the second device can be overridden if a detecting event is detected on the second screen of the second device. In this embodiment, the transmit (TX) side is also referred to as a second device, and the receive (RX) side is also referred to a first device.
  • In addition, optional steps of the embodiment include of the following: detecting a motion event on the first device and generating a sensor data, and sending the sensor data to the second device via the network communication for enabling a corresponding operation on the second device, in which the corresponding operation includes changing orientation of the content displayed on the second screen. In the method of the present embodiment, one of a plurality of sensing ranges can be selected for detecting the motion event, and a threshold value can be provided for detecting the motion event on the first device. The method further includes charging capability to charge the first device with the second device or to charge the second device with the first device by detecting a charging connection between the first device and the second device, retrieving remaining battery amounts from the first device and the second device respectively, and providing an option to charge the first device with the second device, or to charge the second device with the first device.
  • In an another embodiment, a method for performing an actual operation on a second device having a second screen from a first device having a first screen in real time further includes calculating a second position information with respect to a second coordinate system of the second screen based on the first position information.
  • In yet another embodiment, the method for performing the actual operation on the second device having the second screen from the first device having the first screen in real time of the embodiment further includes additional steps which include: receiving compressed multimedia content compressed with a compression ratio from the second device via the network communication, and decompressing and displaying the multimedia content on the first screen of the first device. The compression ratio is determined based on a bandwidth of the network communication. In this embodiment, the decompressed multimedia content can be scaled based on a resolution of the first screen.
  • In another aspect, the present invention provides a method for performing an actual operation on a second device having a second touch panel from a first device having a first touch panel in real time, and the first device is coupled to the second device via a network communication.
  • In one embodiment, the method for performing the actual operation on the second device having the second touch panel from the first device having the first touch panel in real time includes steps as follow: receiving a first touch data associated with a first position information with respect to a first coordinate system of the first touch panel; calculating a second position information with respect to a second coordinate system of the second touch panel based on the first position information; and performing the actual operation on the second device based on the second position information. In this embodiment, the transmit (TX) side is also referred to as a second device, and the receive (RX) side is also referred to a first device. In this embodiment, the performing of the actual operation is overridden if a detecting event is detected on the second touch panel of the second device. The method of this embodiment may further include steps for compressing and decompressing of multimedia content from the second device to the first device as follow: compressing a multimedia content displayed on the second touch panel, and sending the compressed multimedia content from the second device to the first device for decompressing and displaying on the first touch panel via the network communication. This embodiment may further include steps for performing operations on the second device in response to sensor data indicating a motion event on the first device as follow: receiving a sensor data from the first device, the sensor data indicating a motion event on the first device, and performing a corresponding operation on the second device in response to the sensor data. The corresponding operation may include changing orientation of the multimedia content displayed on the second display.
  • In another aspect, the present invention provides a second device configured to be controlled by a first device with a first touch panel, and the first device is coupled to the second device by a network communication.
  • In one embodiment, the second device includes a second touch panel, a compression unit for compressing multimedia content with a compression ratio, a transceiver for transmitting the compressed multimedia content to the first device for decompressing and displaying on the first touch panel and for receiving a touch data associated with a first position information with respect to a first coordinate system of the first touch panel, a coordinate calculating unit for calculating a second position information with respect to a second coordinate system of the second touch panel, and a processor for performing an actual operation on the second device according to the touch data and the second position information. In the second device, the actual operation can be overridden if the second touch panel detects a detecting event or if the second device receives an incoming call.
  • The compression ratio is determined based on a bandwidth of the network communication.
  • In the above embodiments, the transmit (TX) side is also referred to as a second device, and the receive (RX) side is also referred to a first device; in addition, the first touch data is referred to the touch data being formed after converting of the at least one touch signal into touch data.
  • This summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used to limit the scope of the claimed subject matter.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • Embodiments of the present invention will now be described, by way of example, with reference to the accompanying drawings, in which:
  • FIG. 1A is a schematic diagram showing wireless control from a RX side over a TX side according to an embodiment of the present invention.
  • FIG. 1B and FIG. 1C are schematic diagrams showing touch control operations of from the RX side over the TX side shown in FIG. 1A according to an embodiment of the present invention.
  • FIG. 2 is a schematic diagram showing a compression ratio in relation to an available bandwidth between the RX side and the TX side according to an alteration of the present invention.
  • FIG. 3A and FIG. 3B are schematic diagrams showing an image orientation on the TX side in response to a motion detection on the RX side where FIG. 3B is further provided with a switch element for adjusting motion sensitivity according to another alteration of the present invention.
  • FIG. 3C is a schematic diagram showing motion control operations of the RX side over the TX side corresponding to FIGS. 3A and 3B according to an embodiment of the present invention.
  • FIG. 3D and FIG. 3E are schematic diagrams showing adjustment of motion sensitivity on the RX side implemented by an application program on the TX side according to a further alteration of the present invention.
  • FIG. 3F and FIG. 3G are schematic diagrams showing adjustment of magnitude and/or threshold values of application programs on the TX side after motion data is received according to a further alteration of the present invention.
  • FIG. 4 is a flow diagram showing a process for controlling multimedia contents in realtime fashion according to an embodiment of the present invention.
  • FIG. 5 is a flow diagram showing a process for compressing and decompressing multimedia contents in realtime fashion according to an embodiment of the present invention.
  • FIG. 6 is a flow diagram showing another process for compressing and decompressing multimedia contents in realtime fashion according to an embodiment of the present invention.
  • FIG. 7A and FIG. 7B are schematic diagrams showing battery charging between the TX side and the RX side according to a further alteration of the present invention.
  • FIG. 8 is a flow diagram of a battery charging process according to an embodiment of the present invention.
  • FIG. 9 is a flow diagram of a battery charging process according to another embodiment of the present invention.
  • DETAILED DESCRIPTION
  • Refer to FIG. 1A, which is a schematic diagram of a RX side 10 (e.g. pad-like device having a touch-control display) and a TX side 12 (e.g. phone-like device having a computing power & operating system therein) according to an embodiment of the present invention. Advantageously, the RX side 10 is provided for controlling in realtime fashion multimedia contents that are originated from the TX side 12. The multimedia contents can include audio and video streams and be processed to present at the TX side 12 (such as cellphone, smartphone, PDA, and GPS navigation, or any other portable wireless device) having a phone-like screen 1202 and the TX side 12 is coupled to the RX side 10 via a network communication NC, which is preferably a wireless communication and a two-way communication. Noticeably, the multimedia contents may include video and audio data, i.e. the video data can be displayed on the screen 1202 while the audio data can be played via a speaker (not shown) in the TX side 12. Besides, the RX side 10 may be advantageously provided in a CPU-less (without central processing unit) and/or GPU-less (without graphical processing unit) environment for the sake of less power consumption and/or simplified circuit complexity.
  • As shown in FIG. 1A, the TX side 12 may be a computing system or mobile device equipped with CPU and/or GPU computing power, such as a PC, laptop, tablet PC, mobile phone, smart-phone, and may include a processor (not shown) for executing an application program (not shown) of an operation system (OS) 1204 installed in a storage medium (not shown) and a transceiver 1206 coupled to the processor and capable of transmitting and receiving the multimedia contents generated by the application program of OS 1204 executed by the processor. The OS 1204 may be any one of existing or known OS, such as the abovementioned Windows®, iOS®, Android®, Linux®, Blackberry®, Symbian®, Palm®, installed on the TX side 12 such that an application program may be performed or run thereon. The network communication NC between the RX side 10 and the TX side 12 may preferably be any one of the network communication standards or profiles of: IEEE 802.11, IEEE 802.16, CDMA2000, WCDMA, TD-SCDMA, LTE, TDLTE, Bluetooth, Zigbee and combination thereof, or wired communication such as ADSL network, Optical/Fiber network or Cable network.
  • The RX side 10 (refer back to FIG. 1A) includes a pad-like screen 1002 (e.g. touch panel), a coordinate mapping unit 1004 and a transceiver 1006. The screen 1002 is provided for detecting at least one touch signal, and is preferably larger than the screen 1202. The coordinate mapping unit 1004 can be controlled and/or controllable by an operating system (not shown). The coordinate mapping unit 1004 is provided for converting the at least one touch signal (from the screen 1002) into touch data associated with a first position information that indicates a virtual operation on the first screen 1002 corresponding to an actual operation on the second screen 1202, wherein the first position information is with respect to a first coordinate system of the first screen. For example, referred to FIG. 1B, the first position information represents a current cursor position (right side of FIG. 1B) on the pad-like screen 1202 moved by the user's finger (not shown), and thus the mirrored cursor on the phone-like screen 1202 will be moved to the corresponding position (left side of FIG. 1B) such that the user can do any operation on the phone-like screen by touching the pad-like screen 1202 while without touching the phone-like screen 1202. The transceiver 1006 is provided for transmitting the touch data (e.g. the first position information indicative of the current position of cursor on screen 1002) sent from the coordinate mapping unit 1004 to the TX side 12 via the network communication NC. On the other hand, the TX side 12 further includes a coordinate calculating unit 1208 for calculating a second position information (e.g. new cursor position on phone-like screen) with respect to a second coordinate system of the screen 1202 based on the first position information. For example, when the user “points and clicks” the new object on the touch screen 1002 (RX), and thus the same new object on the touch screen 1202 (TX) will be “pointed and clicked” simultaneously because an absolute position (x′, y′) of the same new object on the touch screen 1002 can be obtained as following expression:
      • the mapping absolute position (x′, y′) on the phone-like screen=the current position (x, y)×mapping_ratio,
        where
      • mapping_ratio along x-axis=x-axis_resolution of phone-like screen/x-axis_resolution of pad-like screen;
      • mapping_ratio along y-axis=y-axis_resolution of phone-like screen/y-axis_resolution of pad-like screen;
      • given the same origin for both screens 1202 and 1002.
  • Alternatively, the mapping movement (delta_x′, delta_y′) on the phone-like screen can be obtained as following expression:
      • mapping movement (delta_x′, delta_y′) on the phone-like screen=current movement (delta_x, delta_v) on the pad-like screen×mapping ratio,
        where
      • delta_x=x1−x0; delta_y=y1−y0;
      • mapping_ratio along x-axis=x-axis_resolution of phone-like screen/x-axis resolution of pad-like screen;
      • mapping_ratio along y-axis=y-axis_resolution of phone-like screen/y-axis resolution of pad-like screen;
        given the same origin for both screens 1202 and 1002,
      • where xo and y0 represents the old position on the pad-like screen, and x1 and y1 represents the new position on the pad-like screen while each of the mapping ratios has a value less than one when the display size of the pad-like screen is larger than that of the phone-like screen. Thus, the operating system 1204 can be further provided for performing the actual operation at the second screen based on the second position information, such that the screen 1002 can display results from the actual operation. As a result, the RX side 10 can control multimedia contents played at TX side 12 in realtime fashion due to up-scale and down-scale operations by the coordinate mapping unit 1004 (RX side) and the coordinate calculating unit 1208 (TX side).
  • Noticeably, the first position information may further include the current position of the cursor on the pad-like screen 1002 of the RX side 10, and the resolution data of the RX side 10, e.g. 1920×1080, and the second position information can be calculated so as to obtain the mapping position of the cursor on the phone-like screen 1202, e.g. scaling down to a resolution 480×320. For example, the actual action can be “play a video movie” when the mapping position of the cursor on the screen 1202 is pointed on a “play” icon. As a result, the user can “touch” on the pad-like screen 1002 of the RX side 10 to implement operations without access to the phone-like screen 1202 as if what he/she usually does on the phone-like screen 1202 of the TX side 12. On the other hand, all the tasks requiring extensive computing power/resources are handled by the TX side 12, and only the display task is duplicated (mirrored) to the RX side 10 from the TX side 12 such that the multimedia contents can be presented in realtime fashion between the pad-like screen 1002 of the RX side and the phone-like screen 1202 of the TX side 12.
  • For example, please refer to FIG. 1B and FIG. 1C, which are schematic diagrams of touch control operations of the RX side 10 and the TX side 12 shown in FIG. 1A according to an embodiment of the present invention. As shown in FIG. 1B, a user can do the “open” operation to open a file or folder on the larger screen 1002 of the RX side 10 (right side) and quasi-simultaneously the smaller screen 1202 of the TX side 12 (left side) will mirror the same scene/image as that in the RX side 10, and thus the user can do the “double-click” operation to open some object like file/folder on the larger screen 1002 while utilize computing power from the TX side 12, i.e. the operating system 1204, to run the archive application. As shown in FIG. 1C, a user can do the “key-in” operation on the larger RX side 10 (right side) to type words and simultaneously the smaller screen 1202 of the TX side 12 (left side) will mirror the same scene/image as that in the RX side 10, and thus the user can do the “key-in” operation on the larger screen 1002 while utilize computing power from the TX side 12, i.e. the operating system 1204, to run the text-editing application.
  • In detail, please continue to refer to FIG. 1A, the TX side 12 may further include a compressing unit 1210, for compressing the multimedia contents with a compression ratio, e.g. compressing RGB data into MJPEG data, which requires less time for compression and thus facilitates operating in realtime fashion, and the transceiver 1206 can be further utilized for transmitting the compressed multimedia contents to the screen 1002 for displaying. Specifically, the transceiver 1006 receives the compressed multimedia contents from the transceiver 1206 via the network communication NC, and then compressed multimedia contents are decompressed by a decompressing unit 1010 with the compression ratio, such that the video data of the multimedia contents can be displayed on the screen 1002 while the audio data of the multimedia contents can be played by a speaker 1008. As a result, the multimedia contents can be displayed and controlled in realtime fashion between the screen 1002 of the RX side 10 and the screen 1202 of the TX side 12.
  • Moreover, please refer to FIG. 2, which is a schematic diagram of the RX side 10 and the TX side 12 according to an alteration of the present invention. In the embodiment shown in FIG. 2, the transceiver 1206 can further detect a bandwidth of the network communication NC, and the TX side 12 further includes a dynamic compression unit 1212, for determining the compression ratio of the multimedia content based on the bandwidth of the network communication NC, such that the compressing unit 1210 can compress the multimedia contents at the compression ratio, which is determined by the bandwidth of the network communication NC. Correspondingly, the transceiver 1006 also can detect the bandwidth of the network communication NC, and then the decompressing unit 1010 decompresses the compressed multimedia contents by the compression ratio provided by a bandwidth calculating unit 1012. As a result, the multimedia contents can be compressed with a compression ratio according to a bandwidth of the network communication NC, such that the multimedia contents can be smoothly displayed in realtime fashion between the screen 1002 of the RX side 10 and the screen 1202 of the TX side 12 by facilitating the realtime codec function (e.g. the video compression ratio is determined by the actual available bandwidth). Besides, the transceiver 1006 (RX side) also can directly receive the compression ratio from the dynamic compression unit 1212 (TX side) rather than the bandwidth calculating unit 1012, and then to pass the compression ratio into the decompressing unit 1010 so that the bandwidth calculating unit 1012 can be removed as well.
  • Besides, the present invention discloses another method for controlling in realtime fashion multimedia contents between a RX side having a first touch panel and a TX side having a second touch panel via a network communication. The multimedia contents comprise at least a screen frame data (not shown) where the screen frame data includes the screen resolution data, image data displayed on the screen and audio data. The method comprising the steps of:
      • Step 1: detecting at least one touch signal from the second touch panel 1202, for example, the user may touch the touch-sensitive phone-like screen 1202 to do something, and the “touch” can be detected and a touch signal will be sent to OS for particular operation;
      • Step 2: in response to detecting the touch signal from the second touch panel, compressing the screen frame data on the TX side and transmitting to the RX side so that the RX side decompresses the compressed screen frame data so as to obtain a TX's resolution data, where the TX's resolution data is provided together with a RX's resolution data to scale the screen frame data so as to display the screen frame data to fit with the RX's touch panel;
      • Step 3: detecting at least one touch signal from the first touch panel;
      • Step 4: in response to detecting the touch signal from the first touch panel, the further steps including:
      • Step 4-1: converting the touch signal on the RX side into touch data associated with a first position information defining a virtual operation on the first screen corresponding to an actual operation on the second screen, the first position information being with respect to a first coordinate system of the first screen;
      • Step 4-2: transmitting the touch data from the RX side to the TX side via the network communication, and calculating on the TX side a second position information with respect to a second coordinate system of the second screen based on the first position information, and
      • Step 4-3: performing the actual operation at the second screen based on the second position information; and
      • Step 5: implementing an override mechanism to assign the TX side a higher priority overriding the RX side when the two detecting events (respectively from the TX and RX sides) occur simultaneously, for example, the phone-like screen 1202 will override the pad-like screen 1002 to take control when the user receives an incoming call from the phone-like screen 1202, and the override mechanism is implemented by a software program installed at the TX side. That means the TX side can dominant the RX side to facilitate the communication since the user has paid attention on doing the task at the TX side rather the RX side (e.g. pick-up the incoming call).
  • Furthermore, please refer to FIGS. 3A-3B, which are schematic diagrams of the TX side 12 and the RX side 10 according to another alteration of the present invention. In one embodiment shown in FIG. 3B, the RX side 10 may further include a motion sensor 1014 (including at least one of an accelerometer, a gyroscope and a magnetic compass) for detecting at least one motion event. The receiver 1006 transmits the sensor data (RX) from the motion sensor 1014 after the motion sensor 1014 has detected the motion event on the RX side, and then the sensor data (RX) is sent to the TX side 12 via the network communication NC. A sensor decision unit (TX) may disable sensor data (TX) generated from a motion sensor 1216 on the TX side 12 only if the detecting exists (i.e. the sensor data (RX) from the RX side has being received on the TX side), and then send the sensor data (RX) to a motion process unit 1214. The motion process unit 1214 can convert the sensor data (RX) into motion data (indicating a translation representative of linear acceleration in three-axis directions, and/or an angular rotation representative of Yaw, Roll and Pitch in three-axis directions) on the TX side 12, so as to present the multimedia contents on the screen 1202 and subsequently on the screen 1002 in a same particular orientation (e.g. portrait orientation or landscape orientation). The particular orientation is determined by the operating system 1204 according to the motion data (RX) such that the image at the screen 1202 will be presented at the landscape or portrait orientation. As a result, the multimedia contents can be displayed and controlled in realtime fashion between the screen 1002 of the RX side 10 and the screen 1202 of the TX side 12 through motion operations.
  • For example, please refer to FIG. 3C, which is a schematic diagrams of motion control operations between the RX side 10 and the TX side 12 shown in FIGS. 3A-3B according to one embodiment of the present invention. As shown in the right side of FIG. 3C, a user can do the “game play” driving motion, e.g. moving the image (RX) of the larger pad-like screen 1002 from Landscape to Portrait, and thus the mirrored image (TX) on the smaller phone-like screen 1002 (shown on left side of FIG. 3C) is quasi-simultaneously adjusted from Landscape to Portrait on the TX side 12 performed by the operating system 1204 based on the motion data converted from the sensor data (RX) generated by the motion sensor 1014 (RX). Thus, the user can play the game on the pad-like screen 1002 while utilize computing power from the TX side 12, i.e. the operating system 1204, to run the game application actually on the phone-like screen 1202.
  • Noticeably, the motion sensor 1014 may have a plurality of sensing ranges, i.e. different sensing sensitivities, or a particular sensing range for generating the sensor data. When the motion sensor 1014 has a plurality of sensing ranges, one of the multiple sensing ranges can be selected by a control code determined by a switch signal so as to enable a motion chip to generate one motion data based on the selected sensing range determined by the control code.
  • In a first embodiment, the motion sensor 1014 has a plurality of sensing ranges. As shown in FIG. 3A-3B, the RX side 10 detects whether a switch signal is invoked by a switch element 1016 operative to a user, and the switch element 1016 may issue an adjustment signal to a motion chip (i.e. motion sensor 1014) after the switch signal was invoked (e.g. user turns the switch element 1016 on). Then, the motion sensor 1014 selects one of the plurality of sensing ranges of the motion chip based on the adjustment signal, and generates the motion data by the motion chip based on the selected sensing range so as to enable an interaction between the RX side 10 and the TX side 12. For example, the motion chip has a sensing range of ±1 G at its “off” (original) state and ±2 G at “on” state. When the switch element 1016 is turned to “on” from “off” by the user, the motion chip may have a new sensing range of ±2 G, where G is defined as 9.81 m/s2 for gravitational acceleration.
  • In a second embodiment, the motion sensor 1014 has a plurality of sensing ranges. As shown in FIG. 3D-3E, the TX side 12 detects whether a switch signal is invoked by an application program 1220 (e.g. game software/program), and an adjustment signal is generated in response to the invoked switch signal. Then, the motion sensor 1014 selects one of the plurality of sensing ranges of the motion chip of the motion sensor 1014 based on the adjustment signal, and generates the motion data by the motion chip based on the selected sensing range so as to enable an interaction between the RX side 10 and the TX side 12. For example, the game software may provide an input user interface to allow the user to choose one of three options for selecting particular sensing range among the plurality of sensing ranges where Option1 stands for ±1 G; Option2 for ±2 G; and Option3 for ±3 G, respectively. A particular switch signal will be sent out to the RX side after the selected Option2 is chosen by the user, for instance. The motion sensor 1014 (RX) is requested to change its sensing range to ±2 G directly by the switch signal, or by an adjusted signal in response to the switch signal.
  • In a third embodiment, the motion sensor 1014 has a particular sensing range for generating the sensor data. As shown in FIG. 3F-3G, the TX side 12 detects whether a switch signal is invoked by an application program 1222 of the operating system 1204, and an adjustment signal is sent in response to the invoked switch signal. Then, the TX side 12 may provide a user interface to allow the user to adjust a threshold value of the application program 1222 (e.g. game software/program) based on either the adjustment signal or the invoked switch signal such that the threshold value of the application program 1222 can be increased at a ratio corresponding to the adjustment signal or the invoked switch signal. Subsequently, the motion data may be generated based on sensor data that is sent from the motion sensor 1014 (on the TX side) triggered by the switch signal, so as to enable an interaction between the RX side 10 and the TX side 12.
  • In a fourth embodiment, the motion sensor 1014 has a particular sensing range for generating the sensor data. As shown in FIG. 3D, the TX side 12 detects whether a switch signal is invoked by the application program 1222 of the operating system 1204, and issues an adjustment signal in response to the invoked switch signal. Then, the TX side 12 may provide a user interface to allow the user to adjust a magnitude of the motion data by means of increasing the magnitude's value at a ratio based on either the adjustment signal or the switch signal, so as to enable an interaction between the RX side 10 and the TX side 12. Noticeably, in the above third and fourth embodiment, the TX side 12 can be configured to adjust the threshold value of the application program 1222 and the magnitude of the motion data, respectively, but the TX side 12 can also simultaneously adjust the threshold value of the application program 1222 and the matmitude of the motion data, while retaining both merits.
  • Noticeably the spirit of the present invention is to utilize a RX side to control multimedia contents played at transmitting entities of different types of operating systems in realtime fashion, wherein the RX side only transmits touch data and sensor data for a TX side to perform actual operation, and displays the multimedia contents of the TX side. Those skilled in the art should make modifications or alterations accordingly. For example, the RX side 10 can be foldable, portable, or any other types to meet each user's need(s).
  • Besides, the network communication NC may preferably include a Wi-Fi communication and a Bluetooth (BT) communication, and since the touch data and sensor data have less data volume while the multimedia contents have greater data volume, the touch data and the motion date are preferably transmitted to the TX side 12 via the BT communication, while the multimedia contents are preferably transmitted to the RX side 10 via the Wi-Fi communication. It can be understood that the transceiver 1006 and 1206 of the present invention may be a wireless module adapting other communication standards or profiles including such as IEEE 802.16 (WiMAX), CDMA2000, WCDMA, TD-SCDMA, LTE, TDLTE, Bluetooth-profile and combination thereof or a wired module.
  • Furthermore, the video data of the multimedia contents decompressed and outputted by the decompressing unit 1010 may preferably from a MJPEG format to a RGB format and transmitted or outputted to the screen 1002 in a LVDS format via a display interface converter such as a RGB-to-LVDS converter. It can be understood that other video stream formats such as ITDMI and related interfaces may also be possible. On the other hand, the audio data of the multimedia contents processed and outputted may preferably be transmitted via an audio interface such as an I2S interface or standard and further to the speaker 1008.
  • Moreover, it is preferable that the multimedia contents transmitted and processed in video streaming may be made “real-time” or without significant delays with respect to time for the sake of low latency. In other words, in a preferred explanatory example of the present invention recited herein, the multimedia contents may be communicated between the RX side 10 and the TX side 12 and processed to output “real-time” video and audio streams to the screen 1002 and/or the speaker 1008 of the RX side 10 of the present invention. Preferably, the multimedia contents may be processed in video streaming (i.e. encoding and decoding) and communicated there between as data packets and the duration or delay of time involved in a “real-time” processing and transmission may be less than 100 ms for the sake of low latency; in one example, it may preferably be less than 10 ms, and in another example, it may approximately be 20 ms.
  • Operations of the RX side 10 and the TX side 12 shown in FIG IA can be summarized into a controlling process 40 shown in FIG. 4. The controlling process 40 includes the following steps:
      • Step 400: Detect at least one touch signal via the screen 1002, and convert the at least one touch signal into touch data associated with a first position information defining a virtual operation on the screen 1002 corresponding to an actual operation on the screen 1202, the first position information being with respect to a first coordinate system of the screen 1002.
      • Step 402: Transmit the touch data to the TX side 12 via the network communication NC, and calculate a second position information with respect to a second coordinate system of the screen 1202 based on the first position information.
      • Step 404: Perform the actual operation at the screen 1202 based on the second position information, and display results from the actual operation onto the screen 1002 so as to control the multimedia contents in realtime fashion between the screen 1002 of the RX side 10 and the screen 1202 of the TX side 12.
      • Step 406: End.
  • Operations of the RX side 10 shown in FIG. 1A can be summarized into a receiving process 50 shown in FIG. 5. The receiving process 50 includes the following steps:
      • Step 500: Receive compressed multimedia contents from the TX side 12.
      • Step 502: Convert the multimedia contents for decompression.
      • Step 504: Scale the decompressed data based on a resolution of the RX side 10.
      • Step 506: Output displayable multimedia contents to the RX side 10.
      • Step 508: Detect whether there is touch signal from the RX side 10. If yes, go to step 510; otherwise go to step 500.
      • Step 510: Output touch data to the TX side 12. Go to step 500.
  • Operations of the RX side 10 and the TX side 12 shown in FIG. 2A can be summarized into a receiving and transmitting process 60 shown in FIG. 6. The receiving and transmitting process 60 includes the following steps:
      • Step 600: Receive compressed multimedia contents from the TX side 12.
      • Step 602: Determine a compression ratio based on a bandwidth of the network communication NC.
      • Step 604: Convert the multimedia contents based on the compression ratio.
      • Step 606: Decompress the compressed data, and scale thereof if needed.
      • Step 608: Output displayable multimedia contents to the RX side 10.
      • Step 610: Detect whether there is touch signal from the RX side 10. If yes, go to step 510; otherwise go to step 600.
      • Step 612: Output touch data to obtain position data.
      • Step 614: Perform actions based on the position data at the TX side 12. Go to step 600.
  • Detail descriptions of the controlling process 40, the receiving processes 50 and the receiving and transmitting process 60 can be derived by referring to the above description, and are not narrated hereinafter.
  • On the other hand, please refer to FIG. 7A-7B, which is a schematic diagram of the RX side 10 according to another alteration of the present invention. As shown in FIG. 7A-7B, the TX side 12 further comprises a charging application 1224, charging detector 1226 and battery cells 1228, and the RX side 10 further comprises a charging controller 1018 and battery cells 1020.
  • In one embodiment, system default charging operations of the TX side 12 and the RX side 10 shown in FIG. 7A-7B can be summarized into charging process 80 as shown in FIG. 8. The charging process 80 includes the following steps:
      • Step 800: Detect an event of a charging connection between the TX side 12 and the RX side 10. The charging connection is advantageously provided for conforming to the USB standard. It is advantageously applicable to the two-way charging process.
      • Step 802: Retrieve remaining battery amounts (Rt and Rr) respectively from battery cells 1228 on the TX side 12 and battery cells 1020 on the RX side 10.
      • Step 804: Compare the two remaining battery amounts (Rt and Rr) respectively with two predetermined battery capacity percentages (Tt and Tr) by a power management mechanism that is executed by either the charging application 1224 in the operating system or the charging controller 1018. If each of the remaining battery amounts is below the predetermined battery capacity percentage (i.e. Rt<Tt and Rr<Tr), go to step 806; if the remaining battery amount of the RX side 10 is higher than its predetermined battery capacity percentage while the remaining battery amount of the RX side 10 is higher than that of the TX side 12 (i.e. Rr>Tr and Rr>Rt), go to step 810; if the remaining battery amount of the TX side 12 is higher than its predetermined battery capacity percentage while the remaining battery amount of the TX side 12 is higher than that of the RX side 10 (i.e. Rt>Tt and Rt>Rr), go to step 814.
      • Step 806: Send a first message of recharging information shown as “Please Use Power Adaptor to charge Pad and Cellphone”.
      • Step 808: Charging operation 1: Use Power Adaptor to charge Pad and Cellphone.
      • Step 810: Send a second message of recharging information shown as “Please Use Pad to charge Cellphone”.
      • Step 812: Charging operation 2: Use Pad to charge Cellphone.
      • Step 814: Send a third message of recharging information “Please Use Cellphone to charge Pad”.
      • Step 816: Charging operation 3: Use Cellphone to charge Pad.
  • In another embodiment, user selection charging operations of the TX side 12 and the RX side 10 shown in FIG. 7A-7B can be summarized into charging process 90 as shown in FIG. 9. The charging process 90 includes the following steps:
      • Step 900: Detect an event of a charging connection between the TX side 12 and the RX side 10. The charging connection is advantageously provided for conforming to the USB standard. It is also applicable to the two-way charging process.
      • Step 902: Retrieve remaining battery amounts (Rt and Rr) respectively from battery cells 1228 on the TX side 12 and battery cells 1020 on the RX side 10.
      • Step 904: Present a user interface to allow a user's selection of battery charging.
      • Step 905: If a user chooses option 1 for charging Cellphone, go to step 906; if the user chooses option 2 for charging Pad, go to step 908.
      • Step 906: Use Pad to charge Cellphone.
      • Step 908: Use Cellphone to charge Pad.
      • Step 910: Compare the two remaining battery amounts with two predetermined battery capacity percentages (Tt and Tr) by a power management mechanism that is executed by either the charging application 1224 in the operating system or the charging controller 1018. If each of the remaining battery amounts is below the predetermined battery capacity percentage (i.e. Rt<Tt and Rr<Tr), go to step 912; if the remaining battery amount of the RX side 10 is higher than its predetermined battery capacity percentage, and simultaneously the remaining battery amount of the RX side 10 is higher than that of the TX side 12 (i.e. Rr>Tr and Rr>Rt), go to step 916; if the remaining battery amount of the TX side 12 is higher than its predetermined battery capacity percentage, and simultaneously the remaining battery amount of the TX side 12 is higher than that of the RX side 10 (i.e. Rt>Tt and Rt>Rr), go to step 920.
      • Step 912: Send a first message of recharging information shown as “Please Use Power Adaptor to charge Pad and Cellphone”.
      • Step 914: Charging operation 1: Use Power Adaptor to charge Pad and Cellphone. Come to END.
      • Step 916: Send a second message of recharging information shown as “Please Use Pad to charge Cellphone”.
      • Step 918: Charging operation 2: Use Pad to charge Cellphone. Come to END.
      • Step 920: Send a third message of recharging information “Please Use Cellphone to charge Pad”.
      • Step 922: Charging operation 3: Use Cellphone to charge Pad. Come to END.
  • The event of a charging connection can be detected either by a charging detector 1226 on the TX side 12 or by a charging controller 1018 on the RX side 10. The charging controller 1018 can be implemented by a power management chip. The charging detector 1226 can receive the remaining battery capacity of the battery cells 1228 on the TX side 12; so as to send the battery capacity information to the user interface by means of the charging application 1224 managed by OS 1204, and thus users can view the icon that indicates the actual measurement of the TX's battery capacity. The charging controller 1218 on the RX side 10 can receive the actual measurement of the RX's battery capacity, and sends the measurement to the charging detector 1226 on the TX side 12, so as to facilitate the charging detector 1226 to determine what condition can be meet to implement the appropriate charging operation.
  • For example, when each of the remaining battery amounts is below the predetermined battery capacity percentage (i.e. Rt<Tt and Rr<Tr), a power adaptor is used to charge both the RX side 10 and the TX side 12; when the remaining battery amount of the RX side 10 is higher than its predetermined battery capacity percentage, and simultaneously the remaining battery amount of the RX side 10 is higher than that of the TX side 12 (i.e. Rr>Tr and Rr>Rt), the RX side 10 is used to charge the TX side 12; when the remaining battery amount of the TX side 12 is higher than its predetermined battery capacity percentage, and simultaneously the remaining battery amount of the TX side 12 is higher than that of the RX side 10 (i.e. Rt>Tt and Rt>Rr), the TX side 12 is used to charge the RX side 10. It is noted that the RX side may be compulsorily switched to the power saving mode by the override mechanism when the user receives an incoming call from the TX side, that is, the user currently uses the TX side rather the RX side to either answer or dial the call. Therefore, it makes sense for the user to temporarily keep the RX side idle while he or she is busy for phone call.
  • In the prior art, a universal apparatus capable of communicating with the any example of new breed of electronic devices running any certain type of operating systems is lack. In comparison, the present invention utilizes a RX side to control multimedia contents played at transmitting entities of different types of operating systems in realtime fashion, wherein the RX side only transmits touch data and sensor data for a TX side to perform actual operation, and displays the multimedia contents of the TX side. Furthermore, the present invention can compress multimedia contents a bandwidth of network communication, such that the multimedia contents can be smoothly displayed in realtime fashion between the screen 1002 of the RX side 10 and the screen 1202 of the TX side 12. Besides, the TX side 12 can also be configured as a server having encoding hardware and the RX side 10 as a client having decoding hardware adapted for cloud computing environment running video streaming for the purpose of low latency requirement. Those skilled in the art will readily observe that numerous modifications and alterations of the device and method may be made while retaining the teachings of the invention. Accordingly, the above disclosure should be construed as limited only by the metes and bounds of the appended claims.

Claims (20)

What is claimed is:
1. A method for performing an actual operation on a second device having a second screen from a first device having a first screen in real time, the first device is coupled to the second device via a network communication, the first screen is configured to display a mirrored content of the second screen, the first screen includes touch panel capability, the method comprises:
detecting a first touch signal on the first screen of the first device;
converting the first touch signal into a first touch data associated with a first position information with respect to a first coordinate system of the first screen; and
transmitting the first touch data to the second device via the network communication, so that the actual operation is performed on the second device.
2. The method of claim 1, further comprising:
calculating a second position information with respect to a second coordinate system of the second screen based on the first position information.
3. The method of claim 1, further comprising:
receiving compressed multimedia content compressed with a compression ratio from the second device via the network communication; and
decompressing and displaying the multimedia content on the first screen of the first device.
4. The method of claim 3, wherein the compression ratio is determined based on a bandwidth of the network communication.
5. The method of claim 3, further comprising:
scaling the decompressed multimedia content based on a resolution of the first screen.
6. The method of claim 1, wherein the actual operation performed on the second device is overridden if a detecting event is detected on the second screen of the second device.
7. The method of claim 1, further comprising:
detecting a motion event on the first device and generating a sensor data; and
sending the sensor data to the second device via the network communication for enabling a corresponding operation on the second device.
8. The method of claim 7, wherein the corresponding operation includes changing orientation of the content displayed on the second screen.
9. The method of claim 7, further comprising:
selecting one of a plurality of sensing ranges for detecting the motion event.
10. The method of claim 7, further comprising:
providing a threshold value for detecting the motion event on the first device.
11. The method of claim 1, further comprising:
detecting a charging connection between the first device and the second device;
retrieving remaining battery amounts from the first device and the second device respectively;
providing an option to charge the first device with the second device, or to charge the second device with the first device.
12. A method for performing an actual operation on a second device having a second touch panel from a first device having a first touch panel in real time, the first device is coupled to the second device via a network communication, the method comprises:
receiving a first touch data associated with a first position information with respect to a first coordinate system of the first touch panel;
calculating a second position information with respect to a second coordinate system of the second touch panel based on the first position information; and
performing the actual operation on the second device based on the second position information.
13. The method of claim 12, further comprising:
compressing a multimedia content displayed on the second touch panel; and
sending the compressed multimedia content from the second device to the first device for decompressing and displaying on the first touch panel via the network communication.
14. The method of claim 12, wherein the performing of the actual operation is overridden if a detecting event is detected on the second touch panel of the second device.
15. The method of claim 12, further comprising:
receiving a sensor data from the first device, the sensor data indicating a motion event on the first device; and
performing a corresponding operation on the second device in response to the sensor data.
16. The method of claim 15, wherein the corresponding operation includes changing orientation of the multimedia content displayed on the second display.
17. A second device configured to be controlled by a first device with a first touch panel, the first device is coupled to the second device by a network communication, the second device comprising:
a second touch panel;
a compression unit for compressing multimedia content with a compression ratio;
a transceiver for transmitting the compressed multimedia content to the first device for decompressing and displaying on the first touch panel, and for receiving a first touch data associated with a first position information with respect to a first coordinate system of the first touch panel;
a coordinate calculating unit for calculating a second position information with respect to a second coordinate system of the second touch panel; and
a processor for performing an actual operation on the second device according to the first touch data and the second position information.
18. The second device of claim 17, wherein the actual operation is overridden if the second touch panel detects a detecting event.
19. The second device of claim 17, wherein the actual operation is overridden if the second device receives an incoming call.
20. The second device of claim 17, wherein the compression ratio is determined based on a bandwidth of the network communication.
US14/069,384 2010-11-05 2013-11-01 Method and apparatus for controlling contents in realtime fashion Abandoned US20140055401A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US14/069,384 US20140055401A1 (en) 2010-11-05 2013-11-01 Method and apparatus for controlling contents in realtime fashion

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US41037110P 2010-11-05 2010-11-05
US13/184,584 US8605048B2 (en) 2010-11-05 2011-07-18 Method and apparatus for controlling multimedia contents in realtime fashion
US14/069,384 US20140055401A1 (en) 2010-11-05 2013-11-01 Method and apparatus for controlling contents in realtime fashion

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US13/184,584 Continuation US8605048B2 (en) 2010-11-05 2011-07-18 Method and apparatus for controlling multimedia contents in realtime fashion

Publications (1)

Publication Number Publication Date
US20140055401A1 true US20140055401A1 (en) 2014-02-27

Family

ID=46019153

Family Applications (3)

Application Number Title Priority Date Filing Date
US13/184,584 Expired - Fee Related US8605048B2 (en) 2010-11-05 2011-07-18 Method and apparatus for controlling multimedia contents in realtime fashion
US13/253,993 Abandoned US20120114047A1 (en) 2010-11-05 2011-10-06 Wireless communication system for transmitting high resolution multimedia data and method thereof
US14/069,384 Abandoned US20140055401A1 (en) 2010-11-05 2013-11-01 Method and apparatus for controlling contents in realtime fashion

Family Applications Before (2)

Application Number Title Priority Date Filing Date
US13/184,584 Expired - Fee Related US8605048B2 (en) 2010-11-05 2011-07-18 Method and apparatus for controlling multimedia contents in realtime fashion
US13/253,993 Abandoned US20120114047A1 (en) 2010-11-05 2011-10-06 Wireless communication system for transmitting high resolution multimedia data and method thereof

Country Status (2)

Country Link
US (3) US8605048B2 (en)
CN (1) CN202602826U (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140247196A1 (en) * 2010-12-31 2014-09-04 Ebay Inc. Methods and systems for displaying content on multiple network devices with a simple command
US20140282068A1 (en) * 2013-03-15 2014-09-18 SingTel Idea Factory Pte. Ltd. Systems and methods for transferring of objects among mobile devices based on pairing and matching using actions and/or gestures associated with the mobile device
US20150042559A1 (en) * 2013-08-09 2015-02-12 Lenovo (Beijing) Co., Ltd. Information Processing Method And Electronic Device Thereof, Image Calibration Method And Apparatus, And Electronic Device Thereof
US20160313960A1 (en) * 2015-04-24 2016-10-27 Lenovo (Beijing) Co., Ltd. Control Method, Control Apparatus, And Electronic Device
US10171261B2 (en) * 2014-04-24 2019-01-01 Vivint, Inc. Remote talk down to panel, camera and speaker

Families Citing this family (33)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9198084B2 (en) 2006-05-26 2015-11-24 Qualcomm Incorporated Wireless architecture for a traditional wire-based protocol
US9398089B2 (en) 2008-12-11 2016-07-19 Qualcomm Incorporated Dynamic resource sharing among multiple wireless devices
US9264248B2 (en) 2009-07-02 2016-02-16 Qualcomm Incorporated System and method for avoiding and resolving conflicts in a wireless mobile display digital interface multicast environment
US9582238B2 (en) 2009-12-14 2017-02-28 Qualcomm Incorporated Decomposed multi-stream (DMS) techniques for video display systems
US8781000B2 (en) * 2010-12-30 2014-07-15 Vixs Systems, Inc. Dynamic video data compression
US9787725B2 (en) 2011-01-21 2017-10-10 Qualcomm Incorporated User input back channel for wireless displays
US20130013318A1 (en) * 2011-01-21 2013-01-10 Qualcomm Incorporated User input back channel for wireless displays
US9065876B2 (en) 2011-01-21 2015-06-23 Qualcomm Incorporated User input back channel from a wireless sink device to a wireless source device for multi-touch gesture wireless displays
US10135900B2 (en) 2011-01-21 2018-11-20 Qualcomm Incorporated User input back channel for wireless displays
US9413803B2 (en) 2011-01-21 2016-08-09 Qualcomm Incorporated User input back channel for wireless displays
US9503771B2 (en) 2011-02-04 2016-11-22 Qualcomm Incorporated Low latency wireless display for graphics
US10108386B2 (en) 2011-02-04 2018-10-23 Qualcomm Incorporated Content provisioning for wireless back channel
KR20130054579A (en) * 2011-11-17 2013-05-27 삼성전자주식회사 Display apparatus and control method thereof
JP6106983B2 (en) 2011-11-30 2017-04-05 株式会社リコー Image display device, image display system, method and program
US9339691B2 (en) 2012-01-05 2016-05-17 Icon Health & Fitness, Inc. System and method for controlling an exercise device
US9525998B2 (en) 2012-01-06 2016-12-20 Qualcomm Incorporated Wireless display with multiscreen service
KR102058947B1 (en) * 2012-05-24 2019-12-24 엘지전자 주식회사 Mobile terminal and gps engine control method therof
US9632648B2 (en) 2012-07-06 2017-04-25 Lg Electronics Inc. Mobile terminal, image display device and user interface provision method using the same
TWI637312B (en) * 2012-09-19 2018-10-01 三星電子股份有限公司 Method for displaying information on transparent display device, display device therewith, and computer-readable recording medium therefor
JP2014068145A (en) * 2012-09-25 2014-04-17 Kyocera Corp Portable terminal, display control program and display control method
WO2014153158A1 (en) 2013-03-14 2014-09-25 Icon Health & Fitness, Inc. Strength training apparatus with flywheel and related methods
CN105848733B (en) 2013-12-26 2018-02-13 爱康保健健身有限公司 Magnetic resistance mechanism in hawser apparatus
US10433612B2 (en) 2014-03-10 2019-10-08 Icon Health & Fitness, Inc. Pressure sensor to quantify work
WO2015191445A1 (en) 2014-06-09 2015-12-17 Icon Health & Fitness, Inc. Cable system incorporated into a treadmill
WO2015195965A1 (en) 2014-06-20 2015-12-23 Icon Health & Fitness, Inc. Post workout massage device
US9883101B1 (en) * 2014-07-23 2018-01-30 Hoyos Integrity Corporation Providing a real-time via a wireless communication channel associated with a panoramic video capture device
US10391361B2 (en) 2015-02-27 2019-08-27 Icon Health & Fitness, Inc. Simulating real-world terrain on an exercise device
US10272317B2 (en) 2016-03-18 2019-04-30 Icon Health & Fitness, Inc. Lighted pace feature in a treadmill
US10625137B2 (en) 2016-03-18 2020-04-21 Icon Health & Fitness, Inc. Coordinated displays in an exercise device
US10493349B2 (en) 2016-03-18 2019-12-03 Icon Health & Fitness, Inc. Display on exercise device
US10671705B2 (en) 2016-09-28 2020-06-02 Icon Health & Fitness, Inc. Customizing recipe recommendations
KR102447016B1 (en) * 2017-11-01 2022-09-27 삼성디스플레이 주식회사 Display driver integrated circuit, display system, and method for driving display driver integrated circuit
CN113946302B (en) * 2020-07-07 2022-10-25 华为技术有限公司 Method and device for opening file

Citations (29)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3708785A (en) * 1970-07-31 1973-01-02 Searle Medidata Inc Data scanner for real time interfacing of a computer and plural remote units
US5062059A (en) * 1986-12-08 1991-10-29 Sunriver Corporation Apparatus and method for communication between host CPU and remote terminal
US5638426A (en) * 1993-10-12 1997-06-10 Multimedia Systems Corporation Interactive system for a closed cable network
US5721878A (en) * 1995-06-07 1998-02-24 International Business Machines Corporation Multimedia control system and method for controlling multimedia program presentation
US5721815A (en) * 1995-06-07 1998-02-24 International Business Machines Corporation Media-on-demand communication system and method employing direct access storage device
US20030137426A1 (en) * 2001-12-07 2003-07-24 Eric Anthony Early warning near-real-time security system
US20040060063A1 (en) * 2002-09-24 2004-03-25 Russ Samuel H. PVR channel and PVR IPG information
US20040240650A1 (en) * 2003-05-05 2004-12-02 Microsoft Corporation Real-time communications architecture and methods for use with a personal computer system
US20050015803A1 (en) * 2002-11-18 2005-01-20 Macrae Douglas B. Systems and methods for providing real-time services in an interactive television program guide application
US20050021642A1 (en) * 2003-05-27 2005-01-27 Shunichiro Nonaka Method and apparatus for moving image conversion, method and apparatus for moving image transmission, and programs therefor
US20050093868A1 (en) * 2003-10-30 2005-05-05 Microsoft Corporation Distributed sensing techniques for mobile devices
US20050155043A1 (en) * 2004-01-08 2005-07-14 Schulz Kurt S. Human-machine interface system and method for remotely monitoring and controlling a machine
US20060031787A1 (en) * 2003-08-02 2006-02-09 Viswanath Ananth System and method for real-time configurable monitoring and management of task performance systems
US20060041655A1 (en) * 2004-05-06 2006-02-23 Marty Holloway Bi-directional remote control for remotely controllable apparatus
US20060242589A1 (en) * 2005-04-26 2006-10-26 Rod Cooper System and method for remote examination services
US20070035528A1 (en) * 2004-02-10 2007-02-15 Bruce Hodge Method and apparatus for determining and retrieving positional information
US7224382B2 (en) * 2002-04-12 2007-05-29 Image Masters, Inc. Immersive imaging system
US20090019492A1 (en) * 2007-07-11 2009-01-15 United Video Properties, Inc. Systems and methods for mirroring and transcoding media content
US20100064230A1 (en) * 2008-09-09 2010-03-11 Applied Systems, Inc. Method and apparatus for remotely displaying screen files and efficiently handling remote operator input
US20100076993A1 (en) * 2008-09-09 2010-03-25 Applied Systems, Inc. Method and apparatus for remotely displaying a list by determining a quantity of data to send based on the list size and the display control size
US20100073454A1 (en) * 2008-09-17 2010-03-25 Tandberg Telecom As Computer-processor based interface for telepresence system, method and computer program product
US20100179005A1 (en) * 2000-06-16 2010-07-15 Skyhawke Technologies, Llc. Personal golfing assistant and method and system for graphically displaying golf related information and for collection, processing and distribution of golf related data
US20110188760A1 (en) * 2010-02-03 2011-08-04 Oculus Info Inc. System and Method for Creating and Displaying Map Projections related to Real-Time Images
US8208007B2 (en) * 2004-04-21 2012-06-26 Telepresence Technologies, Llc 3-D displays and telepresence systems and methods therefore
US20140002361A1 (en) * 2012-06-28 2014-01-02 Lee E. Ballard Systems and methods for remote mouse pointer management
US20140053189A1 (en) * 2012-08-17 2014-02-20 Electronics And Telecommunications Research Institute Apparatus and method for providing personalized home screen
US20140223490A1 (en) * 2013-02-07 2014-08-07 Shanghai Powermo Information Tech. Co. Ltd. Apparatus and method for intuitive user interaction between multiple devices
US20140380193A1 (en) * 2013-06-24 2014-12-25 Microsoft Corporation Showing interactions as they occur on a whiteboard
US20150163281A1 (en) * 2013-12-06 2015-06-11 Vmware, Inc. Moving objects of a remote desktop in unstable network environments

Family Cites Families (26)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6018765A (en) * 1996-01-23 2000-01-25 Storage Concepts, Inc. Multi-channel multimedia data server
US6970183B1 (en) * 2000-06-14 2005-11-29 E-Watch, Inc. Multimedia surveillance and monitoring system including network configuration
US6203432B1 (en) 1999-05-11 2001-03-20 Madcatz, Inc. System for adjusting the response characteristic of an electronic game input device
US7628730B1 (en) * 1999-07-08 2009-12-08 Icon Ip, Inc. Methods and systems for controlling an exercise apparatus using a USB compatible portable remote device
US6351991B1 (en) 2000-06-05 2002-03-05 Schlumberger Technology Corporation Determining stress parameters of formations from multi-mode velocity data
KR100396892B1 (en) * 2000-10-05 2003-09-03 삼성전자주식회사 Video MP3 system, apparatus and method for generating reduced video data, and apparatus and method for restoring the reduced video data
JP2003046938A (en) * 2001-07-30 2003-02-14 Fujitsu Ltd Semiconductor device
US20030059014A1 (en) * 2001-09-24 2003-03-27 Teleware, Inc. Multi-media communication management system with caller managed hold system
JP2004220133A (en) 2003-01-10 2004-08-05 Toshiba Corp Information processor, and operation control method for information processor
JP2005267257A (en) 2004-03-18 2005-09-29 Nara Institute Of Science & Technology Handwritten information input system
US20050229228A1 (en) * 2004-04-07 2005-10-13 Sandeep Relan Unicast cable content delivery
US8370888B2 (en) * 2004-06-22 2013-02-05 University Of Southern California Hydra: high-performance data recording architecture for streaming media
JP4220943B2 (en) 2004-07-15 2009-02-04 ソフトバンクモバイル株式会社 Electronics
TWI332797B (en) * 2005-06-30 2010-11-01 Realtek Semiconductor Corp Cordless multimedia system and image display system
US7927216B2 (en) 2005-09-15 2011-04-19 Nintendo Co., Ltd. Video game system with wireless modular handheld controller
JP4262726B2 (en) 2005-08-24 2009-05-13 任天堂株式会社 Game controller and game system
US8065710B2 (en) * 2006-03-02 2011-11-22 At& T Intellectual Property I, L.P. Apparatuses and methods for interactive communication concerning multimedia content
CN101110759A (en) * 2006-07-20 2008-01-23 朗迅科技公司 Peer-to-peer file download system for IPTV network
JP3133600U (en) 2007-04-10 2007-07-19 永之助 古閑 Accelerometer type pen-type pointing device.
US7961747B2 (en) * 2007-07-30 2011-06-14 Broadcom Corporation Display device with conversion capability for portable media player
TWI362605B (en) 2007-08-30 2012-04-21 Ind Tech Res Inst Method for adjusting sensing range and sensitivity and inertia interactive apparatus and system using thereof
US8416848B2 (en) * 2007-12-21 2013-04-09 Broadcom Corporation Device adaptive video transmission system for use with layered video coding and methods for use therewith
US20090238263A1 (en) * 2008-03-20 2009-09-24 Pawan Jaggi Flexible field based energy efficient multimedia processor architecture and method
TWM342547U (en) * 2008-05-29 2008-10-11 Grandtec Electronic Corp Transmission switching device between USB and HDMI
US9166633B2 (en) * 2010-01-21 2015-10-20 Qualcomm Incorporated Systems and methods for interfacing a white space device with a host device
US20110276624A1 (en) * 2010-05-07 2011-11-10 Haseeb Akhtar Converged communication with web 2.0 based screen triggers

Patent Citations (29)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3708785A (en) * 1970-07-31 1973-01-02 Searle Medidata Inc Data scanner for real time interfacing of a computer and plural remote units
US5062059A (en) * 1986-12-08 1991-10-29 Sunriver Corporation Apparatus and method for communication between host CPU and remote terminal
US5638426A (en) * 1993-10-12 1997-06-10 Multimedia Systems Corporation Interactive system for a closed cable network
US5721878A (en) * 1995-06-07 1998-02-24 International Business Machines Corporation Multimedia control system and method for controlling multimedia program presentation
US5721815A (en) * 1995-06-07 1998-02-24 International Business Machines Corporation Media-on-demand communication system and method employing direct access storage device
US20100179005A1 (en) * 2000-06-16 2010-07-15 Skyhawke Technologies, Llc. Personal golfing assistant and method and system for graphically displaying golf related information and for collection, processing and distribution of golf related data
US20030137426A1 (en) * 2001-12-07 2003-07-24 Eric Anthony Early warning near-real-time security system
US7224382B2 (en) * 2002-04-12 2007-05-29 Image Masters, Inc. Immersive imaging system
US20040060063A1 (en) * 2002-09-24 2004-03-25 Russ Samuel H. PVR channel and PVR IPG information
US20050015803A1 (en) * 2002-11-18 2005-01-20 Macrae Douglas B. Systems and methods for providing real-time services in an interactive television program guide application
US20040240650A1 (en) * 2003-05-05 2004-12-02 Microsoft Corporation Real-time communications architecture and methods for use with a personal computer system
US20050021642A1 (en) * 2003-05-27 2005-01-27 Shunichiro Nonaka Method and apparatus for moving image conversion, method and apparatus for moving image transmission, and programs therefor
US20060031787A1 (en) * 2003-08-02 2006-02-09 Viswanath Ananth System and method for real-time configurable monitoring and management of task performance systems
US20050093868A1 (en) * 2003-10-30 2005-05-05 Microsoft Corporation Distributed sensing techniques for mobile devices
US20050155043A1 (en) * 2004-01-08 2005-07-14 Schulz Kurt S. Human-machine interface system and method for remotely monitoring and controlling a machine
US20070035528A1 (en) * 2004-02-10 2007-02-15 Bruce Hodge Method and apparatus for determining and retrieving positional information
US8208007B2 (en) * 2004-04-21 2012-06-26 Telepresence Technologies, Llc 3-D displays and telepresence systems and methods therefore
US20060041655A1 (en) * 2004-05-06 2006-02-23 Marty Holloway Bi-directional remote control for remotely controllable apparatus
US20060242589A1 (en) * 2005-04-26 2006-10-26 Rod Cooper System and method for remote examination services
US20090019492A1 (en) * 2007-07-11 2009-01-15 United Video Properties, Inc. Systems and methods for mirroring and transcoding media content
US20100064230A1 (en) * 2008-09-09 2010-03-11 Applied Systems, Inc. Method and apparatus for remotely displaying screen files and efficiently handling remote operator input
US20100076993A1 (en) * 2008-09-09 2010-03-25 Applied Systems, Inc. Method and apparatus for remotely displaying a list by determining a quantity of data to send based on the list size and the display control size
US20100073454A1 (en) * 2008-09-17 2010-03-25 Tandberg Telecom As Computer-processor based interface for telepresence system, method and computer program product
US20110188760A1 (en) * 2010-02-03 2011-08-04 Oculus Info Inc. System and Method for Creating and Displaying Map Projections related to Real-Time Images
US20140002361A1 (en) * 2012-06-28 2014-01-02 Lee E. Ballard Systems and methods for remote mouse pointer management
US20140053189A1 (en) * 2012-08-17 2014-02-20 Electronics And Telecommunications Research Institute Apparatus and method for providing personalized home screen
US20140223490A1 (en) * 2013-02-07 2014-08-07 Shanghai Powermo Information Tech. Co. Ltd. Apparatus and method for intuitive user interaction between multiple devices
US20140380193A1 (en) * 2013-06-24 2014-12-25 Microsoft Corporation Showing interactions as they occur on a whiteboard
US20150163281A1 (en) * 2013-12-06 2015-06-11 Vmware, Inc. Moving objects of a remote desktop in unstable network environments

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180293044A1 (en) * 2010-12-31 2018-10-11 Ebay Inc. Methods and systems for displaying content on multiple networked devices with a simple command
US10007477B2 (en) * 2010-12-31 2018-06-26 Ebay Inc. Methods and systems for displaying content on multiple networked devices with a simple command
US11650781B2 (en) * 2010-12-31 2023-05-16 Ebay Inc. Methods and systems for displaying content on multiple networked devices with a simple command
US9367281B2 (en) * 2010-12-31 2016-06-14 Ebay Inc. Methods and systems for displaying content on multiple network devices with a simple command
US20160259616A1 (en) * 2010-12-31 2016-09-08 Ebay Inc. Methods and systems for displaying content on multiple networked devices with a simple command
US20220147306A1 (en) * 2010-12-31 2022-05-12 Ebay Inc. Methods and systems for displaying content on multiple networked devices with a simple command
US11269583B2 (en) * 2010-12-31 2022-03-08 Ebay Inc. Methods and systems for displaying content on multiple networked devices with a simple command
US10747491B2 (en) * 2010-12-31 2020-08-18 Ebay Inc. Methods and systems for displaying content on multiple networked devices with a simple command
US20140247196A1 (en) * 2010-12-31 2014-09-04 Ebay Inc. Methods and systems for displaying content on multiple network devices with a simple command
US20140282068A1 (en) * 2013-03-15 2014-09-18 SingTel Idea Factory Pte. Ltd. Systems and methods for transferring of objects among mobile devices based on pairing and matching using actions and/or gestures associated with the mobile device
US9691357B2 (en) * 2013-08-09 2017-06-27 Lenovo (Beijing) Co., Ltd. Information processing method and electronic device thereof, image calibration method and apparatus, and electronic device thereof
US20150042559A1 (en) * 2013-08-09 2015-02-12 Lenovo (Beijing) Co., Ltd. Information Processing Method And Electronic Device Thereof, Image Calibration Method And Apparatus, And Electronic Device Thereof
US10171261B2 (en) * 2014-04-24 2019-01-01 Vivint, Inc. Remote talk down to panel, camera and speaker
US10001962B2 (en) * 2015-04-24 2018-06-19 Lenovo (Beijing) Co., Ltd. Control method, control apparatus, and electronic device
US20160313960A1 (en) * 2015-04-24 2016-10-27 Lenovo (Beijing) Co., Ltd. Control Method, Control Apparatus, And Electronic Device

Also Published As

Publication number Publication date
US20120114047A1 (en) 2012-05-10
CN202602826U (en) 2012-12-12
US20120113029A1 (en) 2012-05-10
US8605048B2 (en) 2013-12-10

Similar Documents

Publication Publication Date Title
US8605048B2 (en) Method and apparatus for controlling multimedia contents in realtime fashion
US11693496B2 (en) Display method and device
US20230289131A1 (en) Method and system for dynamically generating different user environments with secondary devices with displays of various form factors
WO2021057830A1 (en) Information processing method and electronic device
US9310613B2 (en) Mobile wireless display for accessing data from a host and method for controlling
US20220201205A1 (en) Video stream processing method, device, terminal device, and computer-readable storage medium
US20110001699A1 (en) Remote control of host application using motion and voice commands
US20220398059A1 (en) Multi-window display method, electronic device, and system
WO2015130793A1 (en) Backward-compatible apparatus and method for providing video with both standard and high dynamic range
WO2018161534A1 (en) Image display method, dual screen terminal and computer readable non-volatile storage medium
WO2016197697A1 (en) Gesture control method, device and system
CN106406924B (en) Control method and device for starting and quitting picture of application program and mobile terminal
WO2021115103A1 (en) Display control method and terminal device
WO2020007116A1 (en) Split-screen window adjustment method and apparatus, storage medium and electronic device
KR20140134088A (en) Method and apparatus for using a electronic device
US20150256789A1 (en) Method, apparatus and terminal for playing multimedia content
WO2018145561A1 (en) Code rate control method, electronic device, and computer-readable storage medium
CN109445589B (en) Multimedia file playing control method and terminal equipment
CN111447598B (en) Interaction method and display device
US9787746B2 (en) Method and apparatus for processing multimedia content on a graphic cloud
US20220294844A1 (en) Information Transmission Method And Electronic Device
EP3704861B1 (en) Networked user interface back channel discovery via wired video connection
WO2021012128A1 (en) Mobile terminal-based image display apparatus and method, medium, and electronic device
CN115473757A (en) Dynamic network card drive management system, method, device and equipment of intelligent terminal
CN111124327A (en) Screen control method, multi-screen terminal and computer readable storage medium

Legal Events

Date Code Title Description
AS Assignment

Owner name: BLUESPACE CORPORATION, SAMOA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:YE, ZHOU;LIU, PEI-CHUAN;HUANG, SAN-YUAN;AND OTHERS;REEL/FRAME:031525/0502

Effective date: 20110715

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION