US20140013239A1 - Data sharing between smart devices - Google Patents
Data sharing between smart devices Download PDFInfo
- Publication number
- US20140013239A1 US20140013239A1 US14/024,427 US201314024427A US2014013239A1 US 20140013239 A1 US20140013239 A1 US 20140013239A1 US 201314024427 A US201314024427 A US 201314024427A US 2014013239 A1 US2014013239 A1 US 2014013239A1
- Authority
- US
- United States
- Prior art keywords
- smart device
- image
- data
- shared data
- smart
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L65/00—Network arrangements, protocols or services for supporting real-time applications in data packet communication
- H04L65/40—Support for services or applications
- H04L65/403—Arrangements for multi-party communication, e.g. for conferences
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L12/00—Data switching networks
- H04L12/28—Data switching networks characterised by path configuration, e.g. LAN [Local Area Networks] or WAN [Wide Area Networks]
- H04L12/2803—Home automation networks
- H04L12/2807—Exchanging configuration information on appliance services in a home automation network
- H04L12/2812—Exchanging configuration information on appliance services in a home automation network describing content present in a home automation network, e.g. audio video content
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L67/00—Network arrangements or protocols for supporting network services or applications
- H04L67/50—Network services
- H04L67/75—Indicating network or usage conditions on the user display
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M1/00—Substation equipment, e.g. for use by subscribers
- H04M1/72—Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
- H04M1/724—User interfaces specially adapted for cordless or mobile telephones
- H04M1/72403—User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
- H04M1/72409—User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality by interfacing with external accessories
- H04M1/72412—User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality by interfacing with external accessories using two-way short-range wireless interfaces
Definitions
- Embodiments of the present disclosure relate to the field of electronics. More particularly, embodiments of the present disclosure relate to data sharing between smart devices in a network.
- Smart devices are devices that are digital, active, and computer networked.
- the smart devices are also user configurable and can operate to some extent autonomously.
- Some of the examples of the smart devices may include a personal computer (PC), a tablet PC, a smart TV, a mobile phone, a laptop computer, etc.
- One embodiment of the present disclosure pertains to a smart device for data sharing with one or more associated smart devices connected via a network
- the smart device comprises a transceiver, a display unit, a memory, and a processor coupled to the transceiver, the display unit, and the memory.
- the processor is configured to generate, to display on the display unit, an image for each one of the associated smart devices within coverage of the network and an image of shared data.
- the processor is also configured to effect a change in a first image for a first smart device of the associated smart devices when the first smart device is targeted as a target device for the data sharing, and communicate the shared data with the first smart device via the network when, on the display unit, the image of the shared data is moved to or from a domain of the first image of the first smart device.
- FIG. 1 illustrates an exemplary network view of a smart device interacting with associated smart devices via a network, according to one embodiment of the present disclosure.
- FIG. 2 illustrates an exemplary view of the smart device in FIG. 1 selecting a first smart device as a target device, according to one embodiment of the present disclosure.
- FIG. 3 illustrates an exemplary view of the smart device in FIG. 1 performing data sharing with the first smart device, according to one embodiment of the present disclosure.
- FIGS. 4-6 illustrate an exemplary view of the smart device in FIG. 1 cancelling data sharing with the first smart device and performing data sharing with a second smart device, according to one embodiment of the present disclosure.
- FIG. 7 illustrates an exemplary view of the smart device in FIG. 1 relinquishing control of the shared data, according to one embodiment of the present disclosure.
- FIGS. 8-13 illustrate an exemplary view of the smart device in FIG. 1 interacting with a target device for sharing moving image data, according to one embodiment of the present disclosure.
- FIGS. 14 and 15 illustrate an exemplary process of the smart device in FIG. 1 for sharing data with multiple smart devices, according to one embodiment of the present disclosure.
- FIGS. 16-18 illustrate an exemplary process of the smart device in FIG. 1 for accessing data in a target device, according to one embodiment of the present disclosure.
- FIG. 19 illustrates an exemplary component view of the smart device in FIG. 1 , according to one embodiment of the present invention.
- FIG. 20 illustrates an exemplary network view of the smart device in FIG. 1 interacting with associated smart devices via a cloud computing network, according to one embodiment of the present disclosure.
- FIG. 21 illustrates a process flow chart of an exemplary method of the smart device in FIG. 1 for sharing data with a target device via a network, according to one embodiment of the present disclosure.
- FIG. 22 illustrates a process flow chart of another exemplary method of the smart device in FIG. 1 for sharing data with a target device via a network, according to one embodiment of the present disclosure.
- FIG. 23 illustrates a process flow chart of yet another exemplary method of the smart device in FIG. 1 for sharing data with a target device via a network, according to one embodiment of the present disclosure.
- a smart device and method thereof that seamlessly share data with associated smart devices via a network are disclosed.
- a number of smart devices may be connected through a network.
- a user of a smart device may want to share data with other available smart devices in a convenient and user friendly manner.
- it is often a concern to the user that the time taken to perform the data sharing is too long especially when the size of the data is rather large.
- the present disclosure provides a user interface, method, and device which provide an allure to the user that the data sharing is taking place faster than it actually is.
- images of the associated smart devices within the coverage of the network and an image of shared data (or data to be transmitted) are displayed on the display unit of the smart device.
- the image of the targeted smart device changes its size or complexion (e.g., texture, color, etc.) to indicate that it is being targeted.
- a transmitted image of the shared data e.g., a thumbnail image
- the transmission of the shared data is performed.
- the image of shared data e.g., the thumbnail image of shared data
- the targeted smart device may be informed of the ongoing data sharing.
- the shared data is of large size (e.g., moving image data)
- the transmission of the thumbnail image at the initiation of the data transfer, while the shared data is being processed by a buffer of the targeted device may provide an allure to the user of the targeted device that the data transfer or sharing is taking place much faster than it actually is.
- images of the associated smart devices within the coverage of the network and an image of the shared data are displayed on the display unit of the smart device.
- an image of the targeted smart device changes its size or complexion to indicate that it is being targeted.
- a list of shared data e.g., items of shared data
- the user may affirm an order of an item from the list of shared data by dragging the item to the domain of the smart device from the domain of the image of the targeted smart device on the display unit of the smart device.
- the transmission of the shared data from the targeted device to the smart device is performed.
- the present disclosure provides a smart device and method thereof for data sharing. More particularly, the smart device according to the present disclosure provides a user-friendly user interface which makes it convenient for the user to select a target device for data sharing and perform the data sharing. Further, the smart device according to the present disclosure offers an allure to the user that the data transfer between the smart device and the targeted device is happening much faster than it actually is.
- FIG. 1 illustrates an exemplary network view of a smart device 104 (e.g., a tablet PC) interacting with associated smart devices via a network, according to one embodiment of the present disclosure.
- the smart device 104 and the associated smart devices may be owned by a single user (e.g., a user 102 ), or they may be owned by different users.
- the smart device 104 is connected with a first smart device 106 A (e.g., a tablet PC operated by a user 108 ), a second smart device 106 B (e.g., a smart TV), a third smart device 106 C (e.g., a laptop), and a fourth smart device 106 D (e.g., a smart phone) via a network.
- a first smart device 106 A e.g., a tablet PC operated by a user 108
- a second smart device 106 B e.g., a smart TV
- a third smart device 106 C e.g., a
- the network may be any combination of a short range network, a local area network (LAN), a home area network, a campus network, a global area network, an enterprise private network, a virtual private network, an internetwork, an Internet, an intranet and extranet, an overlay network, etc.
- LAN local area network
- a home area network a home area network
- campus network a campus network
- global area network a global area network
- enterprise private network a virtual private network
- an internetwork an Internet
- intranet and extranet an overlay network, etc.
- the smart device 104 A is configured to display on a display unit 112 , respective images of the associated smarts devices within coverage of the network and an image of shared data 116 .
- the display unit 112 displays a first image of first smart device 114 A, a second image of second smart device 114 B, a third image of third smart device 114 C, and a fourth image of fourth smart device 114 D in addition to the image of shared data 116 .
- the first image of first smart device 114 A, the second image of second smart device 114 B, the third image of third smart device 114 C, and the fourth image of fourth smart device 114 D may be based on icons taking after the real images of the smart devices.
- first image of first smart device 114 A, the second image of second smart device 114 B, the third image of third smart device 114 C, and the fourth image of fourth smart device 114 D may be pictures of the real images of the smart devices.
- a transceiver of the smart device 104 may be configured to receive location data (e.g., GPS data) of the associated smart devices, where the location may be used to arrange the order of the images of associated smart devices 114 A-D. That is, the images of associated smart devices 114 A-D on the display unit 112 may be generated based on the location data which indicate that the first smart device (e.g., the tablet PC) is located to the rightmost direction, whereas the fourth image of fourth smart device 114 D is located to the leftmost direction.
- location data e.g., GPS data
- FIGS. 2-7 illustrate an exemplary view of the smart device 104 in FIG. 1 interacting with a target device for sharing shared data 302 , according to one embodiment of the present disclosure. More particularly, FIG. 2 illustrates an exemplary view of the smart device 104 in FIG. 1 selecting the first smart device 106 A as a target device, according to one embodiment of the present disclosure.
- the first image of the first smart device 114 A on the display unit 112 is changed when the first smart device 114 A is targeted as a target device for data sharing.
- the first smart device 106 A is designated as the target device when a body of the smart device 104 is directed to the first smart device 106 A as illustrated in FIG. 2 .
- the smart device 104 is equipped with a sensor which is configured to sense the orientation of the smart device 104 being directed to the target device.
- the sensor may be a gyro sensor, an acceleration sensor, a geomagnetic sensor, or other type of sensor.
- the target device is identified based on location data of the target device 106 A and tilt data sensed by the gyro sensor when the first smart device is targeted as the target device.
- the size of the first image of first smart device 114 A or the complexion of the first image of first smart device 114 A is changed. For example, the size of the first image of first smart device 114 A on the display unit 112 is increased to indicate the selection of the first smart device 106 A as the target device.
- the first image of first smart device 114 A is highlighted to indicate the selection.
- FIG. 3 illustrates an exemplary view of the smart device in FIG. 1 performing data sharing with the first smart device 106 A, according to one embodiment of the present disclosure.
- the smart device 104 transmits shared data 302 stored in a memory of the smart device 104 and represented by the image of shared data 116 via the network when, on the display unit 112 , the image of shared data 116 is moved into the domain of the first image of first smart device 114 A.
- the shared data 302 comprises still image data 304 .
- the smart device 104 transmits meta-data of the shared data 302 .
- the meta-data of the shared data 302 comprises position data 306 of the image of shared data 116 with respect to coordinates of the first image of first smart device 114 A on the display unit 112 .
- a transmitted image of shared data 308 is displayed at the bottom left portion of a first display unit 310 on the first smart device 106 A. That is, the position data 306 is processed by the first smart device 106 A to display the transmitted image of shared data 308 on a designated area of the first display unit 310 of the first smart device 106 A.
- the image of shared data 116 displayed on the display unit 112 of the smart device 104 is a thumbnail image of the shared data 302 .
- the thumbnail image of the shared data 302 may be to the first smart device 106 A in addition to the shared data 302 and the position data 306 . Since the size of the thumbnail image is small, the transmission of the thumbnail image of the shared data 302 may be completed prior to the transmission of the shared data 302 . This may be especially true when the size of the shared data 302 is too large to delay its transmission. By quickly transmitting the thumbnail image of the shared data 302 as the transmission of the shared data 302 is in progress, the user 108 may be informed of the transmission of the shared data 302 in progress.
- the user 102 may be informed of the transmission of the shared data 302 in progress by identifying the thumbnail image of the shared data 302 (e.g., the transmitted image of shared data 308 ) appearing on the first display unit 310 of the first smart device 106 A.
- FIGS. 4-6 illustrate an exemplary view of the smart device 104 in FIG. 1 cancelling data sharing with the first smart device 106 A and performing data sharing with a second smart device 106 B, according to one embodiment of the present disclosure.
- an image of shared data in part 402 A is displayed on the first display unit 310 of the first smart device 106 A and an image of shared data in part 402 B is displayed on a display unit 404 of the second smart device 106 B when the image of shared data 116 is extended over the first image of first smart device 114 A and the second image of second smart device 114 B displayed on the display unit 112 of the smart device 104 .
- the shared data 302 and the position data 306 of the image of shared data 116 with respect to coordinates of the image of first smart device 114 A are transmitted to the first smart device 106 A. Accordingly, the image of shared data in part 402 A is displayed on the display unit 310 of the first smart device 106 A. In addition, the shared data 302 and the position data 306 of the image of shared data 116 with respect to coordinates of the image of second smart device 114 B are transmitted to the second smart device 106 B. Accordingly, the image of shared data in part 402 B is displayed on the display unit 404 of the second smart device 106 B.
- the transmission of the shared data 302 to the first smart device 106 A is canceled when the image of shared data 116 is moved out of the domain of the first image of first smart device 114 A on the display unit 112 while the control of the shared data 302 is maintained.
- the portion of the shared data 302 transmitted to the first smart device 106 A is deleted or becomes inaccessible such that the user 108 may not access the portion of the shared data 302 using the first smart device 106 A.
- control of the shared data 302 is maintained when the user 102 continues to maintain a contact with the image of shared data 116 while the image of shared data 116 is moved in and out of the first image of first smart device 114 A displayed on the display unit 112 of the smart device 104 .
- the transmission of the shared data 302 is performed until its completion, and the user 102 may not undo the transmission of the shared data 302 to the first smart device 106 A.
- the shared data 302 is transmitted to the second smart device 106 B when the image of shared data 116 enters the domain of the second image of second smart device 114 B. Then, a transmitted image of shared data 502 is displayed on the second display unit 404 of the second smart device 106 B, and the transmission of the shared data 302 is continued until its completion.
- FIG. 6 illustrates an exemplary view where the transmission of the shared data 302 is completed.
- the image of shared data 116 placed on the second image of second smart device 114 B is no longer seen on the display unit 112 .
- the transmitted image of shared data 502 is displayed on the display unit 404 of the second smart device 106 B.
- the size or complexion of the image of second smart device 114 B is changed to indicate that the second smart device 106 B is currently directed by the smart device 104 as the target device.
- FIG. 7 illustrates an exemplary view of the smart device 104 in FIG. 1 relinquishing control of the shared data 304 , according to one embodiment of the present disclosure.
- the control of the smart device 104 for the shared data 302 may be relinquished once a certain time period (e.g., 3 seconds) lapses since the placement of the image of shared data 116 within the domain of the image of the target device (e.g., the image of first smart device 114 A).
- the control of the shared data 302 may be relinquished when the user 108 affirms the receipt of the shared data 302 by pressing the transmitted image of shared data 308 displayed on the first display unit 310 of the first smart device 106 A, as illustrated in FIG. 7 .
- FIGS. 8-13 illustrate an exemplary view of the smart device in FIG. 1 interacting with a target device for sharing moving image data, according to one embodiment of the present disclosure.
- a user 802 may want to play a movie the user has been watching with a smart device 804 (e.g., a tablet PC) on a bigger screen of a first smart device 812 (e.g., a smart TV).
- the first smart device 812 is playing a movie or executing moving image data 816 , which is being displayed on a display unit 814 of the first smart device 812 .
- the smart device 804 displays on its display unit 806 associated smart devices in their image forms, such as a first image of first smart device 808 A, a second image of second smart device 808 B, a third image of third smart device 808 C, and a fourth image of fourth smart device 808 D. Further, the smart device 804 displays an image of moving image data 810 , which may be a thumbnail image of the movie that the user 802 wants to play on the first smart device 812 .
- a transmitted image of moving image data 906 appears on the display unit 814 of the first smart device 812 . That is, the upper part of the image of moving image data 810 appears on the display unit 814 , where the moving image data 816 which has been playing on the first smart device 812 is paused by then.
- the shared data 902 e.g., moving image data 904 which is represented by the image of moving image data 810
- the shared data 902 is transmitted to the first smart device 812 .
- the transmitted image of moving image data 906 is a thumbnail image of the moving image data 904 .
- the user 802 may feel that the transmission of the moving image data 904 is in progress as the user 802 watches the transmitted image of moving image data 906 appearing on the display unit 814 of the first smart device 812 while the moving image data 904 is being transmitted to and buffered by the first smart device 812 .
- the moving image data 904 is played on the first smart device 812 .
- the moving image data 904 e.g., a movie
- a control of the moving image data 904 is maintained. That is, the moving image data 904 played by the first smart device 812 is controlled by touch buttons displayed on the display unit 806 of the smart device 804 , such as a stop/play button 1002 , a fast forward button 1004 , a rewind button 1006 , etc.
- the moving image data 904 being played on the first smart device 812 is stopped when the stop button 1002 or a pause button is pressed by the user 802 .
- the moving image data 816 which was paused and/or was in the background continues to play as the image of moving image data 810 is brought out of the domain of the first image of first smart device 808 A, as illustrated in FIG. 12 .
- the moving image data 904 automatically continues to play on the display unit 806 of the smart device 804 from the point where it was stopped in FIG. 11 .
- the moving image data 904 displayed on the display unit 806 is paused when the image of moving image data 810 is brought out of the domain of the first image of first smart device 808 A. And, the moving image data 904 continues to play on the display unit 806 of the smart device 804 from the point where it was stopped in FIG. 11 only after the play button 1002 is pressed.
- FIGS. 14 and 15 illustrate an exemplary process of the smart device 104 in FIG. 1 for sharing data with multiple smart devices, according to one embodiment of the present disclosure.
- the first smart device 106 A is selected as the target device when the first smart device 106 A is directed, and an input for selecting the first smart device 106 A (e.g., a double click) is applied to the first image of first smart device 114 A.
- the second image of second smart device 114 B changes its appearance when the second smart device 106 B is targeted as an additional target device for data sharing.
- the second smart device 106 B is selected as the additional target device when an input for selecting the second smart device (e.g., a double click) is applied to the second image of second smart device 114 B.
- the shared data 302 and/or the position data 306 are transmitted to the first smart device 106 A and the second smart device 106 B in a multi-cast manner when the image of shared data 116 (e.g., the still image data 304 ) is moved within the domain of the first image of first smart device 114 A and/or the domain of the second image of second smart device 114 B.
- the image of shared data 116 e.g., the still image data 304
- a flick of the image of shared data 116 toward the domain of the first image of first smart device 114 A and/or the domain of the second image of second smart device 114 B may initiate the transmission of the shared data 302 and/or the position data 306 to the first smart device 106 A and the second smart device 106 B.
- the transmitted image of shared data 308 and the transmitted image of shared data 502 appear on the first display unit 310 of the first smart device 106 A and the second display unit 404 of the second smart device 106 B, respectively.
- the first image of first smart device 114 A and the second image of second smart device 114 B maintain the changed look (e.g., a change in size, a change in complexion, etc.) until an input for relinquishing control of the first smart device 106 A and the second smart device 106 B is applied to the smart device 104 .
- FIGS. 16-18 illustrate an exemplary process of the smart device 104 in FIG. 1 for accessing data in a target device, according to one embodiment of the present disclosure.
- the user 102 directs the smart device 104 toward the first smart device 106 A to select it as the target device. Then, the user 102 may tap (e.g., or some other input method, such as clicking for a prolonged time) the first image of first smart device 114 A displayed on the smart device 104 in order to access data stored and/or provided by the first smart device 106 A.
- the data may be stored in a memory of the first smart device 106 A, or the data may be stored in a remote server that may be accessed via the first smart device 106 A.
- a window 1702 appears on the display unit 112 of the smart device 104 .
- an image of shared data 1704 A, an image of shared data 1704 B, an image of shared data 1704 C, and an image of shared data 1704 D are displayed within the window 1702 .
- An additional list of the data in the window 1702 may be accessed using a scroll bar 1706 A and a scroll bar 1706 B.
- shared data represented by the image of shared data 1704 D from the first smart device 106 A is transmitted by the first smart device 106 A and/or received by the smart device 104 when the image of shared data 1704 D is moved from the window 1702 to the domain of the smart device 104 (e.g., the area of the display unit 112 which is outside the window 1702 ).
- FIG. 19 illustrates an exemplary component view of the smart device 104 in FIG. 1 , according to one embodiment of the present invention.
- the smart device 104 comprises a transceiver 1902 , the display unit 112 , a memory 1906 , and a processor 1908 coupled to the transceiver 1902 , the display unit 112 , and the memory 1906 .
- the processor 1908 is configured to generate, to display on the display unit 112 , images for associated smart devices within the coverage of the network and an image of shared data.
- the processor 1908 is also configured to effect a change in an image for one of the associated smart devices when the associated smart device is targeted as the target device for data sharing.
- the processor 1908 is configured to communicate the shared data with the associated smart device selected as the target device via the network when the image of the shared data is moved to or from a domain of the image of the associated smart device displayed on the smart device 104 .
- the smart device 104 also comprises a sensor 1904 .
- the sensor 1904 may be based on a gyro sensor, an acceleration sensor, a geomagnetic sensor, or other type of sensor.
- the processor 1908 is configured to identify the target device based on location data 1912 of the target device and tilt data 1914 sensed by the gyro sensor when the associated smart device is targeted as the target device.
- the shared data may be stored in the memory 1906 .
- FIG. 20 illustrates an exemplary network view of the smart device 104 in FIG. 1 interacting with associated smart devices 106 A-D via a cloud computing network 2002 , according to one embodiment of the present disclosure.
- the smart device 104 is connected to the first smart device 106 A, the second smart device 106 B, the third smart device 106 C, and the fourth smart device 106 D via the cloud computing network 2002 .
- the shared data 302 is stored in a storage device 2008 coupled to a server 2006 serviced by a cloud computing provider 2004 .
- the transmitted image of shared data 308 appears on the first display unit 310 of the first smart device 106 A.
- the shared data 302 is transmitted to the first smart device 106 A.
- the shared data 302 may be transmitted directly from the smart device 104 , which locally stores the shared data 302 in its memory, to the first smart device 106 A. It is appreciated that the device, process, and system illustrated in FIG. 20 may be applied to the case where shared data available in a target device is accessed by a smart device, as illustrated in FIGS. 16-18 .
- FIG. 21 illustrates a process flow chart of an exemplary method of the smart device 104 in FIG. 1 for sharing data with a target device via a network, according to one embodiment of the present disclosure.
- particular components described with respect to FIGS. 1-20 are referenced as performing the process in FIG. 21 .
- the components that are referred to as performing the process are for illustrative purposes only. In some implementations, other components, or a different combination of components within a system, may instead perform the process in FIG. 21 .
- the images 114 A-D for associated smart devices 106 A-D within the coverage of the network of the smart device 104 is displayed on the display unit 112 of the smart device 104 .
- a change in the first image of first smart device 114 A is effected when the first smart device 106 A is targeted as a target device for data sharing.
- the shared data 302 is communicated to the first smart device 106 A via the network when the image of shared data 116 displayed on the display unit 116 of the smart device 104 is moved to or from the domain of the first image of first smart device 114 A.
- FIG. 21 may be implemented in a form of a machine-readable medium embodying a set of instructions that, when executed by a machine, cause the machine to perform any of the operations disclosed herein.
- FIG. 22 illustrates a process flow chart of another exemplary method of the smart device 104 in FIG. 1 for sharing data with a target device via a network, according to one embodiment of the present disclosure.
- particular components described with respect to FIGS. 1-20 are referenced as performing the process in FIG. 22 .
- the components that are referred to as performing the process are for illustrative purposes only. In some implementations, other components, or a different combination of components within a system, may instead perform the process in FIG. 22 .
- the images 114 A-D for associated smart devices 106 A-D within the coverage of the network of the smart device 104 is displayed on the display unit 112 of the smart device 104 .
- a change in the first image of first smart device 114 A is effected when the first smart device 106 A is targeted as a target device for data sharing.
- the shared data 302 is transmitted to the first smart device 106 A via the network when the image of shared data 116 displayed on the display unit 116 of the smart device 104 is moved within the domain of the first image of first smart device 114 A.
- FIG. 22 may be implemented in a form of a machine-readable medium embodying a set of instructions that, when executed by a machine, cause the machine to perform any of the operations disclosed herein.
- FIG. 23 illustrates a process flow chart of yet another exemplary method of the smart device 104 in FIG. 1 for sharing data with a target device via a network, according to one embodiment of the present disclosure.
- FIGS. 1-20 are referenced as performing the process in FIG. 23 . It should be noted that the components that are referred to as performing the process are for illustrative purposes only. In some implementations, other components, or a different combination of components within a system, may instead perform the process in FIG. 23 .
- the images 114 A-D for associated smart devices 106 A-D within the coverage of the network of the smart device 104 is displayed on the display unit 112 of the smart device 104 .
- a change in the first image of first smart device 114 A is effected when the first smart device 106 A is targeted as a target device for data sharing.
- shared data available in the first smart device 106 A is displayed when the first image for first smart device 114 A is accessed.
- operation 2308 as illustrated in FIG.
- the shared data is received from the first smart device 106 A through the network when, on the display unit 112 of the smart device 104 , the image of shared data 116 is moved from the domain of the first image of first smart device 114 A to the domain of the smart device 104 .
- FIG. 23 may be implemented in a form of a machine-readable medium embodying a set of instructions that, when executed by a machine, cause the machine to perform any of the operations disclosed herein.
Abstract
Data sharing between networked smart devices is discussed. One embodiment of the present disclosure pertains to a method for data sharing with one or more associated smart devices via a network. The method comprises displaying, on a display unit of the smart device, an image for each one of the associated smart devices within coverage of the network and an image of shared data. The method also comprises effecting a change in a first image for a first smart device of the associated smart devices when the first smart device is targeted as a target device for the data sharing. The method further comprises communicating the shared data with the first smart device via the network when, on the display unit, the image of the shared data is moved to or from a domain of the first image of the first smart device.
Description
- This application is a Continuation of co-pending application No. 13/061,408 filed Feb. 28, 2011, which is the national phase of PCT International Application No. PCT/KR2011/000484 filed on Jan. 24, 2011. The entire contents of all the above applications are hereby incorporated by reference.
- 1. Field of the Invention
- Embodiments of the present disclosure relate to the field of electronics. More particularly, embodiments of the present disclosure relate to data sharing between smart devices in a network.
- 2. Discussion of Background Art
- Smart devices are devices that are digital, active, and computer networked. The smart devices are also user configurable and can operate to some extent autonomously. Some of the examples of the smart devices may include a personal computer (PC), a tablet PC, a smart TV, a mobile phone, a laptop computer, etc. Recently, there are ongoing efforts to find more effective ways to share data (e.g., audio data, video data, etc.) among the smart devices connected through the network.
- One embodiment of the present disclosure pertains to a smart device for data sharing with one or more associated smart devices connected via a network, where the smart device comprises a transceiver, a display unit, a memory, and a processor coupled to the transceiver, the display unit, and the memory. The processor is configured to generate, to display on the display unit, an image for each one of the associated smart devices within coverage of the network and an image of shared data. The processor is also configured to effect a change in a first image for a first smart device of the associated smart devices when the first smart device is targeted as a target device for the data sharing, and communicate the shared data with the first smart device via the network when, on the display unit, the image of the shared data is moved to or from a domain of the first image of the first smart device.
- Example embodiments are illustrated by way of example and not limitation in the figures of the accompanying drawings, in which like references indicate similar elements and in which:
-
FIG. 1 illustrates an exemplary network view of a smart device interacting with associated smart devices via a network, according to one embodiment of the present disclosure. -
FIG. 2 illustrates an exemplary view of the smart device inFIG. 1 selecting a first smart device as a target device, according to one embodiment of the present disclosure. -
FIG. 3 illustrates an exemplary view of the smart device inFIG. 1 performing data sharing with the first smart device, according to one embodiment of the present disclosure. -
FIGS. 4-6 illustrate an exemplary view of the smart device inFIG. 1 cancelling data sharing with the first smart device and performing data sharing with a second smart device, according to one embodiment of the present disclosure. -
FIG. 7 illustrates an exemplary view of the smart device inFIG. 1 relinquishing control of the shared data, according to one embodiment of the present disclosure. -
FIGS. 8-13 illustrate an exemplary view of the smart device inFIG. 1 interacting with a target device for sharing moving image data, according to one embodiment of the present disclosure. -
FIGS. 14 and 15 illustrate an exemplary process of the smart device inFIG. 1 for sharing data with multiple smart devices, according to one embodiment of the present disclosure. -
FIGS. 16-18 illustrate an exemplary process of the smart device inFIG. 1 for accessing data in a target device, according to one embodiment of the present disclosure. -
FIG. 19 illustrates an exemplary component view of the smart device inFIG. 1 , according to one embodiment of the present invention. -
FIG. 20 illustrates an exemplary network view of the smart device inFIG. 1 interacting with associated smart devices via a cloud computing network, according to one embodiment of the present disclosure. -
FIG. 21 illustrates a process flow chart of an exemplary method of the smart device inFIG. 1 for sharing data with a target device via a network, according to one embodiment of the present disclosure. -
FIG. 22 illustrates a process flow chart of another exemplary method of the smart device inFIG. 1 for sharing data with a target device via a network, according to one embodiment of the present disclosure. -
FIG. 23 illustrates a process flow chart of yet another exemplary method of the smart device inFIG. 1 for sharing data with a target device via a network, according to one embodiment of the present disclosure. - Other features of the present embodiments will be apparent from the accompanying drawings and from the detailed description that follows.
- A smart device and method thereof that seamlessly share data with associated smart devices via a network are disclosed. According to embodiments of this disclosure, a number of smart devices may be connected through a network. A user of a smart device may want to share data with other available smart devices in a convenient and user friendly manner. In addition, it is often a concern to the user that the time taken to perform the data sharing is too long especially when the size of the data is rather large.
- As a solution, the present disclosure provides a user interface, method, and device which provide an allure to the user that the data sharing is taking place faster than it actually is. In one embodiment, images of the associated smart devices within the coverage of the network and an image of shared data (or data to be transmitted) are displayed on the display unit of the smart device. Then, as the user directs the body of the smart device to one of the associated smart devices, the image of the targeted smart device changes its size or complexion (e.g., texture, color, etc.) to indicate that it is being targeted. As the user moves the image of shared data within the domain of the image of the targeted smart device displayed on the smart device, a transmitted image of the shared data (e.g., a thumbnail image) appears on the display unit of the targeted device.
- Further, the transmission of the shared data is performed. In this embodiment, through forwarding the image of shared data (e.g., the thumbnail image of shared data) to the targeted device and displaying the transmitted image of shared data on the display unit of the targeted device, upon the initiation of the data transmission, the user operating the targeted smart device may be informed of the ongoing data sharing. Further, when the shared data is of large size (e.g., moving image data), the transmission of the thumbnail image at the initiation of the data transfer, while the shared data is being processed by a buffer of the targeted device, may provide an allure to the user of the targeted device that the data transfer or sharing is taking place much faster than it actually is.
- In another embodiment, images of the associated smart devices within the coverage of the network and an image of the shared data are displayed on the display unit of the smart device. Then, as the user directs the body of the smart device to one of the associated smart devices, an image of the targeted smart device changes its size or complexion to indicate that it is being targeted. As the user accesses the image of the targeted smart device displayed on the smart device, a list of shared data (e.g., items of shared data) offered by the targeted smart device appears on the display unit of the smart device. The user may affirm an order of an item from the list of shared data by dragging the item to the domain of the smart device from the domain of the image of the targeted smart device on the display unit of the smart device. As a result, the transmission of the shared data from the targeted device to the smart device is performed.
- As described above, the present disclosure provides a smart device and method thereof for data sharing. More particularly, the smart device according to the present disclosure provides a user-friendly user interface which makes it convenient for the user to select a target device for data sharing and perform the data sharing. Further, the smart device according to the present disclosure offers an allure to the user that the data transfer between the smart device and the targeted device is happening much faster than it actually is.
- Reference will now be made in detail to the embodiments of the disclosure, examples of which are illustrated in the accompanying drawings. While the disclosure will be described in conjunction with the embodiments, it will be understood that they are not intended to limit the disclosure to these embodiments. On the contrary, the disclosure is intended to cover alternatives, modifications and equivalents, which may be included within the spirit and scope of the disclosure. Furthermore, in the detailed description, numerous specific details are set forth in order to provide a thorough understanding of the present disclosure. However, it will be obvious to one of ordinary skill in the art that the present disclosure may be practiced without these specific details. In other instances, well known methods, procedures, components, and circuits have not been described in detail as not to unnecessarily obscure aspects of the present disclosure.
-
FIG. 1 illustrates an exemplary network view of a smart device 104 (e.g., a tablet PC) interacting with associated smart devices via a network, according to one embodiment of the present disclosure. Thesmart device 104 and the associated smart devices may be owned by a single user (e.g., a user 102), or they may be owned by different users. InFIG. 1 , thesmart device 104 is connected with a firstsmart device 106A (e.g., a tablet PC operated by a user 108), a secondsmart device 106B (e.g., a smart TV), a thirdsmart device 106C (e.g., a laptop), and a fourthsmart device 106D (e.g., a smart phone) via a network. In one embodiment, the network may be any combination of a short range network, a local area network (LAN), a home area network, a campus network, a global area network, an enterprise private network, a virtual private network, an internetwork, an Internet, an intranet and extranet, an overlay network, etc. - As illustrated in
FIG. 1 , the smart device 104A is configured to display on adisplay unit 112, respective images of the associated smarts devices within coverage of the network and an image of shareddata 116. In the exemplary view ofFIG. 1 , thedisplay unit 112 displays a first image of firstsmart device 114A, a second image of secondsmart device 114B, a third image of thirdsmart device 114C, and a fourth image of fourthsmart device 114D in addition to the image of shareddata 116. The first image of firstsmart device 114A, the second image of secondsmart device 114B, the third image of thirdsmart device 114C, and the fourth image of fourthsmart device 114D may be based on icons taking after the real images of the smart devices. - Alternatively, the first image of first
smart device 114A, the second image of secondsmart device 114B, the third image of thirdsmart device 114C, and the fourth image of fourthsmart device 114D may be pictures of the real images of the smart devices. A transceiver of thesmart device 104 may be configured to receive location data (e.g., GPS data) of the associated smart devices, where the location may be used to arrange the order of the images of associatedsmart devices 114A-D. That is, the images of associatedsmart devices 114A-D on thedisplay unit 112 may be generated based on the location data which indicate that the first smart device (e.g., the tablet PC) is located to the rightmost direction, whereas the fourth image of fourthsmart device 114D is located to the leftmost direction. -
FIGS. 2-7 illustrate an exemplary view of thesmart device 104 inFIG. 1 interacting with a target device for sharing shareddata 302, according to one embodiment of the present disclosure. More particularly,FIG. 2 illustrates an exemplary view of thesmart device 104 inFIG. 1 selecting the firstsmart device 106A as a target device, according to one embodiment of the present disclosure. In one embodiment, the first image of the firstsmart device 114A on thedisplay unit 112 is changed when the firstsmart device 114A is targeted as a target device for data sharing. For example, the firstsmart device 106A is designated as the target device when a body of thesmart device 104 is directed to the firstsmart device 106A as illustrated inFIG. 2 . - In one example implementation, the
smart device 104 is equipped with a sensor which is configured to sense the orientation of thesmart device 104 being directed to the target device. The sensor may be a gyro sensor, an acceleration sensor, a geomagnetic sensor, or other type of sensor. Then, the target device is identified based on location data of thetarget device 106A and tilt data sensed by the gyro sensor when the first smart device is targeted as the target device. In one embodiment, when the firstsmart device 106A is selected as the target device, the size of the first image of firstsmart device 114A or the complexion of the first image of firstsmart device 114A is changed. For example, the size of the first image of firstsmart device 114A on thedisplay unit 112 is increased to indicate the selection of the firstsmart device 106A as the target device. Alternatively, the first image of firstsmart device 114A is highlighted to indicate the selection. -
FIG. 3 illustrates an exemplary view of the smart device inFIG. 1 performing data sharing with the firstsmart device 106A, according to one embodiment of the present disclosure. InFIG. 3 , thesmart device 104 transmits shareddata 302 stored in a memory of thesmart device 104 and represented by the image of shareddata 116 via the network when, on thedisplay unit 112, the image of shareddata 116 is moved into the domain of the first image of firstsmart device 114A. In one embodiment, the shareddata 302 comprises stillimage data 304. - In addition, the
smart device 104 transmits meta-data of the shareddata 302. In one example embodiment, the meta-data of the shareddata 302 comprisesposition data 306 of the image of shareddata 116 with respect to coordinates of the first image of firstsmart device 114A on thedisplay unit 112. As illustrated inFIG. 3 , as the image of shareddata 116 moves onto the bottom left portion of the first image offirst device 114A, a transmitted image of shareddata 308 is displayed at the bottom left portion of afirst display unit 310 on the firstsmart device 106A. That is, theposition data 306 is processed by the firstsmart device 106A to display the transmitted image of shareddata 308 on a designated area of thefirst display unit 310 of the firstsmart device 106A. - In one embodiment, the image of shared
data 116 displayed on thedisplay unit 112 of thesmart device 104 is a thumbnail image of the shareddata 302. Further, the thumbnail image of the shareddata 302 may be to the firstsmart device 106A in addition to the shareddata 302 and theposition data 306. Since the size of the thumbnail image is small, the transmission of the thumbnail image of the shareddata 302 may be completed prior to the transmission of the shareddata 302. This may be especially true when the size of the shareddata 302 is too large to delay its transmission. By quickly transmitting the thumbnail image of the shareddata 302 as the transmission of the shareddata 302 is in progress, theuser 108 may be informed of the transmission of the shareddata 302 in progress. In case the firstsmart device 106A is owned and operated by theuser 102 of thesmart device 104, theuser 102 may be informed of the transmission of the shareddata 302 in progress by identifying the thumbnail image of the shared data 302 (e.g., the transmitted image of shared data 308) appearing on thefirst display unit 310 of the firstsmart device 106A. -
FIGS. 4-6 illustrate an exemplary view of thesmart device 104 inFIG. 1 cancelling data sharing with the firstsmart device 106A and performing data sharing with a secondsmart device 106B, according to one embodiment of the present disclosure. InFIG. 4 , an image of shared data inpart 402A is displayed on thefirst display unit 310 of the firstsmart device 106A and an image of shared data in part 402B is displayed on adisplay unit 404 of the secondsmart device 106B when the image of shareddata 116 is extended over the first image of firstsmart device 114A and the second image of secondsmart device 114B displayed on thedisplay unit 112 of thesmart device 104. In one embodiment, the shareddata 302 and theposition data 306 of the image of shareddata 116 with respect to coordinates of the image of firstsmart device 114A are transmitted to the firstsmart device 106A. Accordingly, the image of shared data inpart 402A is displayed on thedisplay unit 310 of the firstsmart device 106A. In addition, the shareddata 302 and theposition data 306 of the image of shareddata 116 with respect to coordinates of the image of secondsmart device 114B are transmitted to the secondsmart device 106B. Accordingly, the image of shared data in part 402B is displayed on thedisplay unit 404 of the secondsmart device 106B. - In
FIG. 5 , the transmission of the shareddata 302 to the firstsmart device 106A is canceled when the image of shareddata 116 is moved out of the domain of the first image of firstsmart device 114A on thedisplay unit 112 while the control of the shareddata 302 is maintained. When the transmission to the firstsmart device 106A is canceled, the portion of the shareddata 302 transmitted to the firstsmart device 106A is deleted or becomes inaccessible such that theuser 108 may not access the portion of the shareddata 302 using the firstsmart device 106A. - Further, in one example implementation, the control of the shared
data 302 is maintained when theuser 102 continues to maintain a contact with the image of shareddata 116 while the image of shareddata 116 is moved in and out of the first image of firstsmart device 114A displayed on thedisplay unit 112 of thesmart device 104. However, if theuser 102 releases the control of the image of shareddata 116 while the image of shareddata 116 is within the domain of the first image of firstsmart device 114A displayed on thedisplay unit 112, the transmission of the shareddata 302 is performed until its completion, and theuser 102 may not undo the transmission of the shareddata 302 to the firstsmart device 106A. - Further, in
FIG. 5 , the shareddata 302 is transmitted to the secondsmart device 106B when the image of shareddata 116 enters the domain of the second image of secondsmart device 114B. Then, a transmitted image of shareddata 502 is displayed on thesecond display unit 404 of the secondsmart device 106B, and the transmission of the shareddata 302 is continued until its completion. -
FIG. 6 illustrates an exemplary view where the transmission of the shareddata 302 is completed. As the transmission is completed, the image of shareddata 116 placed on the second image of secondsmart device 114B is no longer seen on thedisplay unit 112. In addition, upon completion of the transmission of the shareddata 302, the transmitted image of shareddata 502 is displayed on thedisplay unit 404 of the secondsmart device 106B. Further, the size or complexion of the image of secondsmart device 114B is changed to indicate that the secondsmart device 106B is currently directed by thesmart device 104 as the target device. -
FIG. 7 illustrates an exemplary view of thesmart device 104 inFIG. 1 relinquishing control of the shareddata 304, according to one embodiment of the present disclosure. The control of thesmart device 104 for the shareddata 302 may be relinquished once a certain time period (e.g., 3 seconds) lapses since the placement of the image of shareddata 116 within the domain of the image of the target device (e.g., the image of firstsmart device 114A). Alternatively, the control of the shareddata 302 may be relinquished when theuser 108 affirms the receipt of the shareddata 302 by pressing the transmitted image of shareddata 308 displayed on thefirst display unit 310 of the firstsmart device 106A, as illustrated inFIG. 7 . -
FIGS. 8-13 illustrate an exemplary view of the smart device inFIG. 1 interacting with a target device for sharing moving image data, according to one embodiment of the present disclosure. InFIG. 8 , auser 802 may want to play a movie the user has been watching with a smart device 804 (e.g., a tablet PC) on a bigger screen of a first smart device 812 (e.g., a smart TV). As illustrated inFIG. 8 , the firstsmart device 812 is playing a movie or executing movingimage data 816, which is being displayed on adisplay unit 814 of the firstsmart device 812. In addition, thesmart device 804 displays on itsdisplay unit 806 associated smart devices in their image forms, such as a first image of firstsmart device 808A, a second image of secondsmart device 808B, a third image of thirdsmart device 808C, and a fourth image of fourthsmart device 808D. Further, thesmart device 804 displays an image of movingimage data 810, which may be a thumbnail image of the movie that theuser 802 wants to play on the firstsmart device 812. - In
FIG. 9 , as the firstsmart device 812 is selected as the target device, and theuser 802 brings an upper part of the image of movingimage data 810 within the domain of the image of firstsmart device 808A, a transmitted image of movingimage data 906 appears on thedisplay unit 814 of the firstsmart device 812. That is, the upper part of the image of movingimage data 810 appears on thedisplay unit 814, where the movingimage data 816 which has been playing on the firstsmart device 812 is paused by then. In addition, the shared data 902 (e.g., movingimage data 904 which is represented by the image of moving image data 810) is transmitted to the firstsmart device 812. In one embodiment, the transmitted image of movingimage data 906 is a thumbnail image of the movingimage data 904. In such a way, theuser 802 may feel that the transmission of the movingimage data 904 is in progress as theuser 802 watches the transmitted image of movingimage data 906 appearing on thedisplay unit 814 of the firstsmart device 812 while the movingimage data 904 is being transmitted to and buffered by the firstsmart device 812. - In
FIG. 10 , as the image of movingimage data 810 is brought within the domain of the first image of firstsmart device 808A, the movingimage data 904 is played on the firstsmart device 812. While the moving image data 904 (e.g., a movie) is being executed, a control of the movingimage data 904 is maintained. That is, the movingimage data 904 played by the firstsmart device 812 is controlled by touch buttons displayed on thedisplay unit 806 of thesmart device 804, such as a stop/play button 1002, afast forward button 1004, arewind button 1006, etc. - In
FIG. 11 , the movingimage data 904 being played on the firstsmart device 812 is stopped when thestop button 1002 or a pause button is pressed by theuser 802. The movingimage data 816 which was paused and/or was in the background continues to play as the image of movingimage data 810 is brought out of the domain of the first image of firstsmart device 808A, as illustrated inFIG. 12 . Then, as illustrated inFIG. 13 , the movingimage data 904 automatically continues to play on thedisplay unit 806 of thesmart device 804 from the point where it was stopped inFIG. 11 . Alternatively, the movingimage data 904 displayed on thedisplay unit 806 is paused when the image of movingimage data 810 is brought out of the domain of the first image of firstsmart device 808A. And, the movingimage data 904 continues to play on thedisplay unit 806 of thesmart device 804 from the point where it was stopped inFIG. 11 only after theplay button 1002 is pressed. -
FIGS. 14 and 15 illustrate an exemplary process of thesmart device 104 inFIG. 1 for sharing data with multiple smart devices, according to one embodiment of the present disclosure. InFIG. 14 , the firstsmart device 106A is selected as the target device when the firstsmart device 106A is directed, and an input for selecting the firstsmart device 106A (e.g., a double click) is applied to the first image of firstsmart device 114A. Then, the second image of secondsmart device 114B changes its appearance when the secondsmart device 106B is targeted as an additional target device for data sharing. Further, the secondsmart device 106B is selected as the additional target device when an input for selecting the second smart device (e.g., a double click) is applied to the second image of secondsmart device 114B. - In
FIG. 15 , the shareddata 302 and/or theposition data 306 are transmitted to the firstsmart device 106A and the secondsmart device 106B in a multi-cast manner when the image of shared data 116 (e.g., the still image data 304) is moved within the domain of the first image of firstsmart device 114A and/or the domain of the second image of secondsmart device 114B. Alternatively, once the target devices are designated, a flick of the image of shareddata 116 toward the domain of the first image of firstsmart device 114A and/or the domain of the second image of secondsmart device 114B may initiate the transmission of the shareddata 302 and/or theposition data 306 to the firstsmart device 106A and the secondsmart device 106B. - In
FIG. 15 , when the image of shareddata 116 is moved within or toward the first image of firstsmart device 114A and the second image of secondsmart device 114B, the transmitted image of shareddata 308 and the transmitted image of shareddata 502 appear on thefirst display unit 310 of the firstsmart device 106A and thesecond display unit 404 of the secondsmart device 106B, respectively. Further, the first image of firstsmart device 114A and the second image of secondsmart device 114B maintain the changed look (e.g., a change in size, a change in complexion, etc.) until an input for relinquishing control of the firstsmart device 106A and the secondsmart device 106B is applied to thesmart device 104. -
FIGS. 16-18 illustrate an exemplary process of thesmart device 104 inFIG. 1 for accessing data in a target device, according to one embodiment of the present disclosure. InFIG. 16 , theuser 102 directs thesmart device 104 toward the firstsmart device 106A to select it as the target device. Then, theuser 102 may tap (e.g., or some other input method, such as clicking for a prolonged time) the first image of firstsmart device 114A displayed on thesmart device 104 in order to access data stored and/or provided by the firstsmart device 106A. In one example implementation, the data may be stored in a memory of the firstsmart device 106A, or the data may be stored in a remote server that may be accessed via the firstsmart device 106A. - Then, upon tapping the first image of first
smart device 114A, awindow 1702 appears on thedisplay unit 112 of thesmart device 104. In one example embodiment, an image of shareddata 1704A, an image of shareddata 1704B, an image of shareddata 1704C, and an image of shareddata 1704D are displayed within thewindow 1702. An additional list of the data in thewindow 1702 may be accessed using ascroll bar 1706A and ascroll bar 1706B. - In
FIG. 18 , shared data represented by the image of shareddata 1704D from the firstsmart device 106A is transmitted by the firstsmart device 106A and/or received by thesmart device 104 when the image of shareddata 1704D is moved from thewindow 1702 to the domain of the smart device 104 (e.g., the area of thedisplay unit 112 which is outside the window 1702). -
FIG. 19 illustrates an exemplary component view of thesmart device 104 inFIG. 1 , according to one embodiment of the present invention. As illustrated inFIG. 19 , thesmart device 104 comprises atransceiver 1902, thedisplay unit 112, amemory 1906, and aprocessor 1908 coupled to thetransceiver 1902, thedisplay unit 112, and thememory 1906. In one embodiment, theprocessor 1908 is configured to generate, to display on thedisplay unit 112, images for associated smart devices within the coverage of the network and an image of shared data. Theprocessor 1908 is also configured to effect a change in an image for one of the associated smart devices when the associated smart device is targeted as the target device for data sharing. Further, theprocessor 1908 is configured to communicate the shared data with the associated smart device selected as the target device via the network when the image of the shared data is moved to or from a domain of the image of the associated smart device displayed on thesmart device 104. - In one embodiment, the
smart device 104 also comprises asensor 1904. Further, thesensor 1904 may be based on a gyro sensor, an acceleration sensor, a geomagnetic sensor, or other type of sensor. Further, in one example implementation, theprocessor 1908 is configured to identify the target device based onlocation data 1912 of the target device andtilt data 1914 sensed by the gyro sensor when the associated smart device is targeted as the target device. In addition, the shared data may be stored in thememory 1906. -
FIG. 20 illustrates an exemplary network view of thesmart device 104 inFIG. 1 interacting with associatedsmart devices 106A-D via acloud computing network 2002, according to one embodiment of the present disclosure. InFIG. 20 , thesmart device 104 is connected to the firstsmart device 106A, the secondsmart device 106B, the thirdsmart device 106C, and the fourthsmart device 106D via thecloud computing network 2002. In addition, the shareddata 302 is stored in astorage device 2008 coupled to aserver 2006 serviced by acloud computing provider 2004. - As illustrated in
FIG. 20 , as the image of shareddata 116 is moved within the domain of the first image of firstsmart device 114A displayed on thesmart device 104, the transmitted image of shareddata 308 appears on thefirst display unit 310 of the firstsmart device 106A. Further, the shareddata 302 is transmitted to the firstsmart device 106A. Alternatively, as illustrated inFIGS. 1-19 , the shareddata 302 may be transmitted directly from thesmart device 104, which locally stores the shareddata 302 in its memory, to the firstsmart device 106A. It is appreciated that the device, process, and system illustrated inFIG. 20 may be applied to the case where shared data available in a target device is accessed by a smart device, as illustrated inFIGS. 16-18 . -
FIG. 21 illustrates a process flow chart of an exemplary method of thesmart device 104 inFIG. 1 for sharing data with a target device via a network, according to one embodiment of the present disclosure. In keeping with the previous examples, particular components described with respect toFIGS. 1-20 are referenced as performing the process inFIG. 21 . It should be noted that the components that are referred to as performing the process are for illustrative purposes only. In some implementations, other components, or a different combination of components within a system, may instead perform the process inFIG. 21 . - In
operation 2102, as illustrated inFIG. 1 , theimages 114A-D for associatedsmart devices 106A-D within the coverage of the network of thesmart device 104 is displayed on thedisplay unit 112 of thesmart device 104. Inoperation 2104, as illustrated inFIG. 2 , a change in the first image of firstsmart device 114A is effected when the firstsmart device 106A is targeted as a target device for data sharing. Inoperation 2106, as illustrated inFIG. 3 , the shareddata 302 is communicated to the firstsmart device 106A via the network when the image of shareddata 116 displayed on thedisplay unit 116 of thesmart device 104 is moved to or from the domain of the first image of firstsmart device 114A. - It is appreciated that the methods disclosed in
FIG. 21 may be implemented in a form of a machine-readable medium embodying a set of instructions that, when executed by a machine, cause the machine to perform any of the operations disclosed herein. -
FIG. 22 illustrates a process flow chart of another exemplary method of thesmart device 104 inFIG. 1 for sharing data with a target device via a network, according to one embodiment of the present disclosure. In keeping with the previous examples, particular components described with respect toFIGS. 1-20 are referenced as performing the process inFIG. 22 . It should be noted that the components that are referred to as performing the process are for illustrative purposes only. In some implementations, other components, or a different combination of components within a system, may instead perform the process inFIG. 22 . - In
operation 2202, as illustrated inFIG. 1 , theimages 114A-D for associatedsmart devices 106A-D within the coverage of the network of thesmart device 104 is displayed on thedisplay unit 112 of thesmart device 104. Inoperation 2204, as illustrated inFIG. 2 , a change in the first image of firstsmart device 114A is effected when the firstsmart device 106A is targeted as a target device for data sharing. Inoperation 2206, as illustrated inFIG. 3 , the shareddata 302 is transmitted to the firstsmart device 106A via the network when the image of shareddata 116 displayed on thedisplay unit 116 of thesmart device 104 is moved within the domain of the first image of firstsmart device 114A. - It is appreciated that the methods disclosed in
FIG. 22 may be implemented in a form of a machine-readable medium embodying a set of instructions that, when executed by a machine, cause the machine to perform any of the operations disclosed herein. -
FIG. 23 illustrates a process flow chart of yet another exemplary method of thesmart device 104 inFIG. 1 for sharing data with a target device via a network, according to one embodiment of the present disclosure. - In keeping with the previous examples, particular components described with respect to
FIGS. 1-20 are referenced as performing the process inFIG. 23 . It should be noted that the components that are referred to as performing the process are for illustrative purposes only. In some implementations, other components, or a different combination of components within a system, may instead perform the process inFIG. 23 . - In
operation 2302, as illustrated inFIG. 16 , theimages 114A-D for associatedsmart devices 106A-D within the coverage of the network of thesmart device 104 is displayed on thedisplay unit 112 of thesmart device 104. Inoperation 2304, as illustrated inFIG. 16 , a change in the first image of firstsmart device 114A is effected when the firstsmart device 106A is targeted as a target device for data sharing. Inoperation 2306, as illustrated inFIG. 17 , shared data available in the firstsmart device 106A is displayed when the first image for firstsmart device 114A is accessed. Further, inoperation 2308, as illustrated inFIG. 18 , the shared data is received from the firstsmart device 106A through the network when, on thedisplay unit 112 of thesmart device 104, the image of shareddata 116 is moved from the domain of the first image of firstsmart device 114A to the domain of thesmart device 104. - It is appreciated that the methods disclosed in
FIG. 23 may be implemented in a form of a machine-readable medium embodying a set of instructions that, when executed by a machine, cause the machine to perform any of the operations disclosed herein. - The previous description of the disclosed embodiments is provided to enable any person skilled in the art to make or use the present disclosure. Various modifications to these embodiments will be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other embodiments without departing from the spirit or scope of the disclosure. Thus, the present disclosure is not intended to be limited to the embodiments shown herein but is to be accorded the widest scope consistent with the principles and features disclosed herein.
Claims (20)
1. A smart device for data sharing with at least one associated smart device connected via a network, the smart device comprising:
a transceiver;
a display;
a memory; and
a processor coupled to the transceiver, the display, and the memory and the processor configured to perform steps of:
displaying on the display an image for each one of the at least one associated smart device within a coverage of the network;
effecting a change in a first image for a first smart device of the at least one associated smart device displayed on the display when the first smart device is targeted for the data sharing by directing a body of the smart device to the first smart device;
displaying an image of shared data provided by the first smart device for the data sharing on the display while the first smart device is targeted for the data sharing; and
receiving the shared data from the first smart device via the network in response to an input applied to the image of the shared data.
2. The smart device of claim 1 , wherein the displaying the image of the shared data comprises displaying the image of the shared data within a domain of the first image of the first smart device.
3. The smart device of claim 2 , wherein the domain of the first image of the first smart device comprises a window displayed on the display, and wherein the image of the shared data is displayed within the window.
4. The smart device of claim 3 , wherein the displaying the image of the shared data further comprises displaying an other image of an other shared data provided by the first smart device for the data sharing within the window.
5. The smart device of claim 4 , wherein the input applied to the image of the shared data comprises a touch input moving the image of the shared data outside the window.
6. The smart device of claim 1 , wherein the image for each one of the at least one associated smart device is an icon or a photographed image.
7. The smart device of claim 1 , wherein the shared data is stored in a memory of the first smart device or a remote server accessed by the first smart device.
8. The smart device of claim 1 , wherein the network is a cloud computing network, and the shared data is stored in a server of the cloud computing network.
9. The smart device of claim 1 , wherein the change in the first image for the first smart device includes a change in a size of the first image or a change in a complexion of the first image.
10. The smart device of claim 1 , wherein the shared data is still image data or moving image data.
11. The smart device of claim 10 , wherein the shared data further comprises metadata of the still image data or the moving image data.
12. The smart device of claim 1 , wherein the transceiver is configured to receive location data of the at least one associated smart device.
13. The smart device of claim 12 , further comprising a gyro sensor.
14. The smart device of claim 13 , wherein the processor is configured to identify the target device based on location data of the target device and tilt data sensed by the gyro sensor when the first smart device is targeted as the target device.
15. The smart device of claim 1 , wherein the receiving of the shared data is canceled when the image of the shared data is moved back to the window while control of the image of the shared data is maintained.
16. The smart device of claim 1 , wherein the shared data is still image data.
17. The smart device of claim 1 , wherein the shared data is moving image data.
18. The smart device of claim 1 , wherein the image of the shared data displayed on the display is a thumbnail image of the shared data.
19. The smart device of claim 18 , wherein the thumbnail image of the shared data is received by the smart device prior to the receiving of the shared data.
20. A method of a smart device for data sharing with at least one associated smart device connected via a network, the method comprising:
displaying on a display of the smart device an image for each one of the at least one associated smart device within a coverage of the network;
effecting a change in a first image for a first smart device of the at least one associated smart device displayed on the display when the first smart device is targeted for the data sharing by directing a body of the smart device to the first smart device;
displaying an image of shared data provided by the first smart device for the data sharing on the display while the first smart device is targeted for the data sharing; and
receiving the shared data from the first smart device via the network in response to an input applied to the image of the shared data.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/024,427 US20140013239A1 (en) | 2011-01-24 | 2013-09-11 | Data sharing between smart devices |
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/061,408 US8554897B2 (en) | 2011-01-24 | 2011-01-24 | Data sharing between smart devices |
PCT/KR2011/000484 WO2012102416A1 (en) | 2011-01-24 | 2011-01-24 | Data sharing between smart devices |
US14/024,427 US20140013239A1 (en) | 2011-01-24 | 2013-09-11 | Data sharing between smart devices |
Related Parent Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/KR2011/000484 Continuation WO2012102416A1 (en) | 2011-01-24 | 2011-01-24 | Data sharing between smart devices |
US13/061,408 Continuation US8554897B2 (en) | 2011-01-24 | 2011-01-24 | Data sharing between smart devices |
Publications (1)
Publication Number | Publication Date |
---|---|
US20140013239A1 true US20140013239A1 (en) | 2014-01-09 |
Family
ID=46544998
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/061,408 Active 2031-09-05 US8554897B2 (en) | 2011-01-24 | 2011-01-24 | Data sharing between smart devices |
US14/024,427 Abandoned US20140013239A1 (en) | 2011-01-24 | 2013-09-11 | Data sharing between smart devices |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/061,408 Active 2031-09-05 US8554897B2 (en) | 2011-01-24 | 2011-01-24 | Data sharing between smart devices |
Country Status (2)
Country | Link |
---|---|
US (2) | US8554897B2 (en) |
WO (1) | WO2012102416A1 (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140282103A1 (en) * | 2013-03-16 | 2014-09-18 | Jerry Alan Crandall | Data sharing |
WO2016003943A1 (en) * | 2014-06-29 | 2016-01-07 | Tetherball Technology, LLC | Automatic wireless grid access sharing and monetization from a mobile device app |
US20200084116A1 (en) * | 2013-11-01 | 2020-03-12 | Google Llc | User-configurable delivery of informational events |
Families Citing this family (33)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8074581B2 (en) | 2007-10-12 | 2011-12-13 | Steelcase Inc. | Conference table assembly |
US20140361954A1 (en) * | 2013-06-07 | 2014-12-11 | Lewis Epstein | Personal control apparatus and method for sharing information in a collaboration workspace |
US10631632B2 (en) | 2008-10-13 | 2020-04-28 | Steelcase Inc. | Egalitarian control apparatus and method for sharing information in a collaborative workspace |
US10884607B1 (en) | 2009-05-29 | 2021-01-05 | Steelcase Inc. | Personal control apparatus and method for sharing information in a collaborative workspace |
KR101770206B1 (en) * | 2011-04-06 | 2017-08-22 | 엘지전자 주식회사 | Mobile terminal and user interface providing method using the same |
US20130027613A1 (en) * | 2011-05-03 | 2013-01-31 | Lg Electronics Inc. | Image display apparatus, portable terminal, and methods for operating the same |
KR102008495B1 (en) * | 2012-02-24 | 2019-08-08 | 삼성전자주식회사 | Method for sharing content and mobile terminal thereof |
KR101919008B1 (en) * | 2012-02-24 | 2018-11-19 | 삼성전자주식회사 | Method for providing information and mobile terminal thereof |
US20140040762A1 (en) * | 2012-08-01 | 2014-02-06 | Google Inc. | Sharing a digital object |
KR102009816B1 (en) * | 2012-08-28 | 2019-08-12 | 삼성전자주식회사 | Screen display method and apparatus |
US20140136985A1 (en) * | 2012-11-12 | 2014-05-15 | Moondrop Entertainment, Llc | Method and system for sharing content |
US11558672B1 (en) * | 2012-11-19 | 2023-01-17 | Cox Communications, Inc. | System for providing new content related to content currently being accessed |
WO2014132250A1 (en) * | 2013-02-26 | 2014-09-04 | Adience SER LTD | Generating user insights from images and other data |
US10425468B2 (en) * | 2013-02-28 | 2019-09-24 | Nokia Technologies Oy | User interface transfer |
KR102181223B1 (en) * | 2013-03-15 | 2020-11-23 | 비데리 인코포레이티드 | Systems and methods for distributing, displaying, viewing, and controlling digital art and imaging |
US9880727B2 (en) * | 2013-05-29 | 2018-01-30 | Microsoft Technology Licensing, Llc | Gesture manipulations for configuring system settings |
KR101491045B1 (en) * | 2013-09-25 | 2015-02-10 | 주식회사 픽스트리 | Apparatus and methdo for sharing contents |
EP3054378B1 (en) * | 2013-10-04 | 2022-11-02 | Sony Group Corporation | Information processing device, information processing method, and program |
KR102210704B1 (en) * | 2013-12-24 | 2021-02-02 | 삼성전자주식회사 | User terminal and control method thereof |
US9398250B2 (en) * | 2014-01-06 | 2016-07-19 | Arun Sobti & Associates, Llc | System and apparatus for smart devices based conferencing |
KR102282498B1 (en) | 2014-05-19 | 2021-07-27 | 삼성전자주식회사 | Method and apparatus for processing an input using a display |
US20170244992A1 (en) * | 2014-10-30 | 2017-08-24 | Sharp Kabushiki Kaisha | Media playback communication |
CN105808089A (en) * | 2014-12-30 | 2016-07-27 | 联想(北京)有限公司 | Data transmission method and first electronic device |
KR20160086128A (en) | 2015-01-09 | 2016-07-19 | 삼성전자주식회사 | Channel display apparatus and method thereof |
US20160291703A1 (en) * | 2015-03-31 | 2016-10-06 | Sony Corporation | Operating system, wearable device, and operation method |
CN104954846B (en) * | 2015-07-27 | 2018-09-18 | 北京京东方多媒体科技有限公司 | Element method of adjustment, equipment and system |
US10795449B2 (en) * | 2015-12-11 | 2020-10-06 | Google Llc | Methods and apparatus using gestures to share private windows in shared virtual environments |
KR102498714B1 (en) * | 2016-05-10 | 2023-02-13 | 삼성전자주식회사 | Electronic device and method for providing content |
US10264213B1 (en) | 2016-12-15 | 2019-04-16 | Steelcase Inc. | Content amplification system and method |
WO2018124689A1 (en) * | 2016-12-29 | 2018-07-05 | Samsung Electronics Co., Ltd. | Managing display of content on one or more secondary device by primary device |
US10057715B1 (en) | 2017-03-29 | 2018-08-21 | Honeywell International Inc. | Systems and methods for selecting an optimal device in a home security or automation system for presenting a notification or alert |
CN110109636B (en) * | 2019-04-28 | 2022-04-05 | 华为技术有限公司 | Screen projection method, electronic device and system |
US20230102708A1 (en) * | 2021-09-30 | 2023-03-30 | Lenovo (Beijing) Limited | Information processing method, electronic device, and storage medium |
Citations (72)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5107443A (en) * | 1988-09-07 | 1992-04-21 | Xerox Corporation | Private regions within a shared workspace |
US20020113778A1 (en) * | 2000-10-25 | 2002-08-22 | Junichi Rekimoto | Data input/output system, data input/output method, and program recording medium |
US6545669B1 (en) * | 1999-03-26 | 2003-04-08 | Husam Kinawi | Object-drag continuity between discontinuous touch-screens |
US20030069931A1 (en) * | 2001-09-28 | 2003-04-10 | Katsuyuki Omura | Conference support apparatus, information processor, teleconference system and computer product |
US20040070608A1 (en) * | 2002-10-10 | 2004-04-15 | International Business Machines Corporation | Apparatus and method for transferring files from one machine to another using adjacent desktop displays in a virtual network |
US20040199514A1 (en) * | 2003-04-02 | 2004-10-07 | Ira Rosenblatt | Techniques for facilitating item sharing |
US20050034082A1 (en) * | 2003-08-09 | 2005-02-10 | Lg Electronics Inc. | Content downloading system and method of mobile terminal |
US20050108297A1 (en) * | 2003-11-17 | 2005-05-19 | Microsoft Corporation | Transfer of user profiles using portable storage devices |
US20050219211A1 (en) * | 2004-03-31 | 2005-10-06 | Kotzin Michael D | Method and apparatus for content management and control |
US20050219223A1 (en) * | 2004-03-31 | 2005-10-06 | Kotzin Michael D | Method and apparatus for determining the context of a device |
US20060053389A1 (en) * | 2004-04-16 | 2006-03-09 | Cascade Basic Research Corp. | Graphical user interface for establishing data sharing relationships |
US20060136828A1 (en) * | 2004-12-16 | 2006-06-22 | Taiga Asano | System and method for sharing display screen between information processing apparatuses |
US20060146765A1 (en) * | 2003-02-19 | 2006-07-06 | Koninklijke Philips Electronics, N.V. | System for ad hoc sharing of content items between portable devices and interaction methods therefor |
US20060184540A1 (en) * | 2004-10-21 | 2006-08-17 | Allen Kung | System and method for managing creative assets via a rich user client interface |
US20060256074A1 (en) * | 2005-05-13 | 2006-11-16 | Robert Bosch Gmbh | Sensor-initiated exchange of information between devices |
US20070060193A1 (en) * | 2005-09-14 | 2007-03-15 | Lg Electronics Inc. | Mobile terminal for sending and receiving contents using messaging service and method thereof |
US20070124685A1 (en) * | 2003-04-17 | 2007-05-31 | Lenovo (Singapore) Pte. Ltd. | Remote support for computer or other electronic device |
US20070146347A1 (en) * | 2005-04-22 | 2007-06-28 | Outland Research, Llc | Flick-gesture interface for handheld computing devices |
US20070157101A1 (en) * | 2006-01-04 | 2007-07-05 | Eric Indiran | Systems and methods for transferring data between computing devices |
US20070275714A1 (en) * | 2006-05-25 | 2007-11-29 | Lg Electronics Inc. | Mobile terminal and method of visual data processing |
US20070288599A1 (en) * | 2006-06-09 | 2007-12-13 | Microsoft Corporation | Dragging and dropping objects between local and remote modules |
US20080222688A1 (en) * | 2007-03-08 | 2008-09-11 | Lg Electronics Inc. | Transmitting data file during call |
US20090017799A1 (en) * | 2007-07-13 | 2009-01-15 | Sony Ericsson Mobile Communications Ab | System, device and method for transmitting a file by use of a throwing gesture to a mobile terminal |
US20090075634A1 (en) * | 2005-06-29 | 2009-03-19 | Microsoft Corporation | Data buddy |
US20090077475A1 (en) * | 2007-09-17 | 2009-03-19 | Areae, Inc. | System for providing virtual spaces with separate places and/or acoustic areas |
US20090106666A1 (en) * | 2007-10-18 | 2009-04-23 | Sony Corporation | File transfer method, file transfer apparatus, and file transfer program |
US20090140986A1 (en) * | 2007-11-30 | 2009-06-04 | Nokia Corporation | Method, apparatus and computer program product for transferring files between devices via drag and drop |
US20090164930A1 (en) * | 2007-12-25 | 2009-06-25 | Ming-Yu Chen | Electronic device capable of transferring object between two display units and controlling method thereof |
US20090217177A1 (en) * | 2008-02-27 | 2009-08-27 | Cisco Technology, Inc. | Multi-party virtual desktop |
US20090235191A1 (en) * | 2008-03-11 | 2009-09-17 | Garbow Zachary A | Method for Accessing a Secondary Virtual Environment from Within a Primary Virtual Environment |
US20090244015A1 (en) * | 2008-03-31 | 2009-10-01 | Sengupta Uttam K | Device, system, and method of wireless transfer of files |
US20090254839A1 (en) * | 2008-04-04 | 2009-10-08 | Microsoft Corporation | Techniques to remotely manage a multimedia conference event |
US20090292999A1 (en) * | 2008-05-21 | 2009-11-26 | Smart Technologies Ulc | Desktop sharing method and system |
US20090300549A1 (en) * | 2008-05-30 | 2009-12-03 | Winston Wang | Relationship-based and context-based user interfaces for exchanging data |
US20090309846A1 (en) * | 2008-06-11 | 2009-12-17 | Marc Trachtenberg | Surface computing collaboration system, method and apparatus |
US7689933B1 (en) * | 2005-11-14 | 2010-03-30 | Adobe Systems Inc. | Methods and apparatus to preview content |
US20100095213A1 (en) * | 2008-10-10 | 2010-04-15 | Raph Koster | System and method for providing virtual spaces for access by users via the web |
US20100122184A1 (en) * | 2008-09-19 | 2010-05-13 | Musigy Usa, Inc. | Method and System for Distributed Computing Interface |
US20100138743A1 (en) * | 2008-11-28 | 2010-06-03 | Pei-Yin Chou | Intuitive file transfer method |
US20100146064A1 (en) * | 2008-12-08 | 2010-06-10 | Electronics And Telecommunications Research Institute | Source apparatus, sink apparatus and method for sharing information thereof |
US20100217837A1 (en) * | 2006-12-29 | 2010-08-26 | Prodea Systems , Inc. | Multi-services application gateway and system employing the same |
US20100257471A1 (en) * | 2009-04-03 | 2010-10-07 | International Business Machines Corporation | Programmatic Information Transfer |
US20100281395A1 (en) * | 2007-09-11 | 2010-11-04 | Smart Internet Technology Crc Pty Ltd | Systems and methods for remote file transfer |
US7843454B1 (en) * | 2007-04-25 | 2010-11-30 | Adobe Systems Incorporated | Animated preview of images |
US20110010656A1 (en) * | 2009-07-13 | 2011-01-13 | Ta Keo Ltd | Apparatus and method for improved user interface |
US20110072383A1 (en) * | 2008-05-27 | 2011-03-24 | Martin Olausson | Work Station In A Process Control System |
US20110072344A1 (en) * | 2009-09-23 | 2011-03-24 | Microsoft Corporation | Computing system with visual clipboard |
US20110078573A1 (en) * | 2009-09-28 | 2011-03-31 | Sony Corporation | Terminal apparatus, server apparatus, display control method, and program |
US20110081923A1 (en) * | 2009-10-02 | 2011-04-07 | Babak Forutanpour | Device movement user interface gestures for file sharing functionality |
US20110099497A1 (en) * | 2008-06-04 | 2011-04-28 | Nec Corporation | Method for enabling a mobile user equipment to drag and drop data objects between distributed applications |
US20110163944A1 (en) * | 2010-01-05 | 2011-07-07 | Apple Inc. | Intuitive, gesture-based communications with physics metaphors |
US20110173328A1 (en) * | 2010-01-13 | 2011-07-14 | Samsung Electronics Co., Ltd. | Electronic device, cloud instance, apparatus, method, and system for combining cloud instances with resources of devices |
US20110175920A1 (en) * | 2010-01-13 | 2011-07-21 | Smart Technologies Ulc | Method for handling and transferring data in an interactive input system, and interactive input system executing the method |
US20110202971A1 (en) * | 2010-02-16 | 2011-08-18 | Google Inc. | Server-Based Data Sharing in Computer Applications |
US20110249024A1 (en) * | 2010-04-09 | 2011-10-13 | Juha Henrik Arrasvuori | Method and apparatus for generating a virtual interactive workspace |
US20110289147A1 (en) * | 2010-05-24 | 2011-11-24 | Styles Andrew G | Direction-Conscious Information Sharing |
US20120017165A1 (en) * | 2010-07-19 | 2012-01-19 | Soasta, Inc. | Apparatus and method for combining / correlating data sets |
US20120023410A1 (en) * | 2010-07-20 | 2012-01-26 | Erik Roth | Computing device and displaying method at the computing device |
US20120030567A1 (en) * | 2010-07-28 | 2012-02-02 | Victor B Michael | System with contextual dashboard and dropboard features |
US20120054637A1 (en) * | 2010-08-27 | 2012-03-01 | Nokia Corporation | Method, apparatus, computer program and user interface |
US20120072853A1 (en) * | 2009-03-05 | 2012-03-22 | Krigstroem Anders | Cooperative Drag and Drop |
US20120088548A1 (en) * | 2010-10-06 | 2012-04-12 | Chanphill Yun | Mobile terminal, display device and controlling method thereof |
US20120096368A1 (en) * | 2010-10-14 | 2012-04-19 | Microsoft Corporation | Cloud-based virtual clipboard |
US20120098733A1 (en) * | 2010-10-26 | 2012-04-26 | Ricoh Company, Ltd. | Screen sharing system, screen sharing method, and storage medium |
US20120278727A1 (en) * | 2011-04-29 | 2012-11-01 | Avaya Inc. | Method and apparatus for allowing drag-and-drop operations across the shared borders of adjacent touch screen-equipped devices |
US20130055143A1 (en) * | 2011-08-31 | 2013-02-28 | Smart Technologies Ulc | Method for manipulating a graphical user interface and interactive input system employing the same |
US20130078990A1 (en) * | 2011-09-22 | 2013-03-28 | Mikyung Kim | Mobile device and method for controlling reproduction of contents in mobile device |
US20130201123A1 (en) * | 2012-02-06 | 2013-08-08 | Lg Electronics Inc. | Mobile terminal and electronic communication method using the same |
US20140009394A1 (en) * | 2012-07-06 | 2014-01-09 | Lg Electronics Inc. | Mobile terminal, image display device and user interface provision method using the same |
US20140033058A1 (en) * | 2012-07-30 | 2014-01-30 | Plantronics, Inc. | Use of multiple mobile devices to control an application |
US20140164966A1 (en) * | 2012-12-06 | 2014-06-12 | Samsung Electronics Co., Ltd. | Display device and method of controlling the same |
US20140365923A1 (en) * | 2013-06-10 | 2014-12-11 | Samsung Electronics Co., Ltd. | Home screen sharing apparatus and method thereof |
Family Cites Families (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7188181B1 (en) | 1999-06-30 | 2007-03-06 | Sun Microsystems, Inc. | Universal session sharing |
US7751347B2 (en) | 2002-04-25 | 2010-07-06 | Azurn Networks, Inc. | Converged conferencing appliance methods for concurrent voice and data conferencing sessions over networks |
US7376696B2 (en) * | 2002-08-27 | 2008-05-20 | Intel Corporation | User interface to facilitate exchanging files among processor-based devices |
US7623892B2 (en) * | 2003-04-02 | 2009-11-24 | Palm, Inc. | System and method for enabling a person to switch use of computing devices |
US20080281919A1 (en) * | 2007-05-09 | 2008-11-13 | University Of Georgia Research Foundation, Inc. | System and Method for Sharing Images |
US8428645B2 (en) * | 2007-06-05 | 2013-04-23 | Bindu Rama Rao | Mobile device capable of sharing SMS messages, email screen display locally with other devices |
WO2009102885A1 (en) * | 2008-02-12 | 2009-08-20 | Phone Through, Inc. | Systems and methods for enabling interactivity among a plurality of devices |
US8112490B2 (en) * | 2008-05-15 | 2012-02-07 | Upton Kevin S | System and method for providing a virtual environment with shared video on demand |
US8386620B2 (en) * | 2009-12-15 | 2013-02-26 | Apple Inc. | Ad hoc networking based on content and location |
-
2011
- 2011-01-24 US US13/061,408 patent/US8554897B2/en active Active
- 2011-01-24 WO PCT/KR2011/000484 patent/WO2012102416A1/en active Application Filing
-
2013
- 2013-09-11 US US14/024,427 patent/US20140013239A1/en not_active Abandoned
Patent Citations (72)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5107443A (en) * | 1988-09-07 | 1992-04-21 | Xerox Corporation | Private regions within a shared workspace |
US6545669B1 (en) * | 1999-03-26 | 2003-04-08 | Husam Kinawi | Object-drag continuity between discontinuous touch-screens |
US20020113778A1 (en) * | 2000-10-25 | 2002-08-22 | Junichi Rekimoto | Data input/output system, data input/output method, and program recording medium |
US20030069931A1 (en) * | 2001-09-28 | 2003-04-10 | Katsuyuki Omura | Conference support apparatus, information processor, teleconference system and computer product |
US20040070608A1 (en) * | 2002-10-10 | 2004-04-15 | International Business Machines Corporation | Apparatus and method for transferring files from one machine to another using adjacent desktop displays in a virtual network |
US20060146765A1 (en) * | 2003-02-19 | 2006-07-06 | Koninklijke Philips Electronics, N.V. | System for ad hoc sharing of content items between portable devices and interaction methods therefor |
US20040199514A1 (en) * | 2003-04-02 | 2004-10-07 | Ira Rosenblatt | Techniques for facilitating item sharing |
US20070124685A1 (en) * | 2003-04-17 | 2007-05-31 | Lenovo (Singapore) Pte. Ltd. | Remote support for computer or other electronic device |
US20050034082A1 (en) * | 2003-08-09 | 2005-02-10 | Lg Electronics Inc. | Content downloading system and method of mobile terminal |
US20050108297A1 (en) * | 2003-11-17 | 2005-05-19 | Microsoft Corporation | Transfer of user profiles using portable storage devices |
US20050219211A1 (en) * | 2004-03-31 | 2005-10-06 | Kotzin Michael D | Method and apparatus for content management and control |
US20050219223A1 (en) * | 2004-03-31 | 2005-10-06 | Kotzin Michael D | Method and apparatus for determining the context of a device |
US20060053389A1 (en) * | 2004-04-16 | 2006-03-09 | Cascade Basic Research Corp. | Graphical user interface for establishing data sharing relationships |
US20060184540A1 (en) * | 2004-10-21 | 2006-08-17 | Allen Kung | System and method for managing creative assets via a rich user client interface |
US20060136828A1 (en) * | 2004-12-16 | 2006-06-22 | Taiga Asano | System and method for sharing display screen between information processing apparatuses |
US20070146347A1 (en) * | 2005-04-22 | 2007-06-28 | Outland Research, Llc | Flick-gesture interface for handheld computing devices |
US20060256074A1 (en) * | 2005-05-13 | 2006-11-16 | Robert Bosch Gmbh | Sensor-initiated exchange of information between devices |
US20090075634A1 (en) * | 2005-06-29 | 2009-03-19 | Microsoft Corporation | Data buddy |
US20070060193A1 (en) * | 2005-09-14 | 2007-03-15 | Lg Electronics Inc. | Mobile terminal for sending and receiving contents using messaging service and method thereof |
US7689933B1 (en) * | 2005-11-14 | 2010-03-30 | Adobe Systems Inc. | Methods and apparatus to preview content |
US20070157101A1 (en) * | 2006-01-04 | 2007-07-05 | Eric Indiran | Systems and methods for transferring data between computing devices |
US20070275714A1 (en) * | 2006-05-25 | 2007-11-29 | Lg Electronics Inc. | Mobile terminal and method of visual data processing |
US20070288599A1 (en) * | 2006-06-09 | 2007-12-13 | Microsoft Corporation | Dragging and dropping objects between local and remote modules |
US20100217837A1 (en) * | 2006-12-29 | 2010-08-26 | Prodea Systems , Inc. | Multi-services application gateway and system employing the same |
US20080222688A1 (en) * | 2007-03-08 | 2008-09-11 | Lg Electronics Inc. | Transmitting data file during call |
US7843454B1 (en) * | 2007-04-25 | 2010-11-30 | Adobe Systems Incorporated | Animated preview of images |
US20090017799A1 (en) * | 2007-07-13 | 2009-01-15 | Sony Ericsson Mobile Communications Ab | System, device and method for transmitting a file by use of a throwing gesture to a mobile terminal |
US20100281395A1 (en) * | 2007-09-11 | 2010-11-04 | Smart Internet Technology Crc Pty Ltd | Systems and methods for remote file transfer |
US20090077475A1 (en) * | 2007-09-17 | 2009-03-19 | Areae, Inc. | System for providing virtual spaces with separate places and/or acoustic areas |
US20090106666A1 (en) * | 2007-10-18 | 2009-04-23 | Sony Corporation | File transfer method, file transfer apparatus, and file transfer program |
US20090140986A1 (en) * | 2007-11-30 | 2009-06-04 | Nokia Corporation | Method, apparatus and computer program product for transferring files between devices via drag and drop |
US20090164930A1 (en) * | 2007-12-25 | 2009-06-25 | Ming-Yu Chen | Electronic device capable of transferring object between two display units and controlling method thereof |
US20090217177A1 (en) * | 2008-02-27 | 2009-08-27 | Cisco Technology, Inc. | Multi-party virtual desktop |
US20090235191A1 (en) * | 2008-03-11 | 2009-09-17 | Garbow Zachary A | Method for Accessing a Secondary Virtual Environment from Within a Primary Virtual Environment |
US20090244015A1 (en) * | 2008-03-31 | 2009-10-01 | Sengupta Uttam K | Device, system, and method of wireless transfer of files |
US20090254839A1 (en) * | 2008-04-04 | 2009-10-08 | Microsoft Corporation | Techniques to remotely manage a multimedia conference event |
US20090292999A1 (en) * | 2008-05-21 | 2009-11-26 | Smart Technologies Ulc | Desktop sharing method and system |
US20110072383A1 (en) * | 2008-05-27 | 2011-03-24 | Martin Olausson | Work Station In A Process Control System |
US20090300549A1 (en) * | 2008-05-30 | 2009-12-03 | Winston Wang | Relationship-based and context-based user interfaces for exchanging data |
US20110099497A1 (en) * | 2008-06-04 | 2011-04-28 | Nec Corporation | Method for enabling a mobile user equipment to drag and drop data objects between distributed applications |
US20090309846A1 (en) * | 2008-06-11 | 2009-12-17 | Marc Trachtenberg | Surface computing collaboration system, method and apparatus |
US20100122184A1 (en) * | 2008-09-19 | 2010-05-13 | Musigy Usa, Inc. | Method and System for Distributed Computing Interface |
US20100095213A1 (en) * | 2008-10-10 | 2010-04-15 | Raph Koster | System and method for providing virtual spaces for access by users via the web |
US20100138743A1 (en) * | 2008-11-28 | 2010-06-03 | Pei-Yin Chou | Intuitive file transfer method |
US20100146064A1 (en) * | 2008-12-08 | 2010-06-10 | Electronics And Telecommunications Research Institute | Source apparatus, sink apparatus and method for sharing information thereof |
US20120072853A1 (en) * | 2009-03-05 | 2012-03-22 | Krigstroem Anders | Cooperative Drag and Drop |
US20100257471A1 (en) * | 2009-04-03 | 2010-10-07 | International Business Machines Corporation | Programmatic Information Transfer |
US20110010656A1 (en) * | 2009-07-13 | 2011-01-13 | Ta Keo Ltd | Apparatus and method for improved user interface |
US20110072344A1 (en) * | 2009-09-23 | 2011-03-24 | Microsoft Corporation | Computing system with visual clipboard |
US20110078573A1 (en) * | 2009-09-28 | 2011-03-31 | Sony Corporation | Terminal apparatus, server apparatus, display control method, and program |
US20110081923A1 (en) * | 2009-10-02 | 2011-04-07 | Babak Forutanpour | Device movement user interface gestures for file sharing functionality |
US20110163944A1 (en) * | 2010-01-05 | 2011-07-07 | Apple Inc. | Intuitive, gesture-based communications with physics metaphors |
US20110173328A1 (en) * | 2010-01-13 | 2011-07-14 | Samsung Electronics Co., Ltd. | Electronic device, cloud instance, apparatus, method, and system for combining cloud instances with resources of devices |
US20110175920A1 (en) * | 2010-01-13 | 2011-07-21 | Smart Technologies Ulc | Method for handling and transferring data in an interactive input system, and interactive input system executing the method |
US20110202971A1 (en) * | 2010-02-16 | 2011-08-18 | Google Inc. | Server-Based Data Sharing in Computer Applications |
US20110249024A1 (en) * | 2010-04-09 | 2011-10-13 | Juha Henrik Arrasvuori | Method and apparatus for generating a virtual interactive workspace |
US20110289147A1 (en) * | 2010-05-24 | 2011-11-24 | Styles Andrew G | Direction-Conscious Information Sharing |
US20120017165A1 (en) * | 2010-07-19 | 2012-01-19 | Soasta, Inc. | Apparatus and method for combining / correlating data sets |
US20120023410A1 (en) * | 2010-07-20 | 2012-01-26 | Erik Roth | Computing device and displaying method at the computing device |
US20120030567A1 (en) * | 2010-07-28 | 2012-02-02 | Victor B Michael | System with contextual dashboard and dropboard features |
US20120054637A1 (en) * | 2010-08-27 | 2012-03-01 | Nokia Corporation | Method, apparatus, computer program and user interface |
US20120088548A1 (en) * | 2010-10-06 | 2012-04-12 | Chanphill Yun | Mobile terminal, display device and controlling method thereof |
US20120096368A1 (en) * | 2010-10-14 | 2012-04-19 | Microsoft Corporation | Cloud-based virtual clipboard |
US20120098733A1 (en) * | 2010-10-26 | 2012-04-26 | Ricoh Company, Ltd. | Screen sharing system, screen sharing method, and storage medium |
US20120278727A1 (en) * | 2011-04-29 | 2012-11-01 | Avaya Inc. | Method and apparatus for allowing drag-and-drop operations across the shared borders of adjacent touch screen-equipped devices |
US20130055143A1 (en) * | 2011-08-31 | 2013-02-28 | Smart Technologies Ulc | Method for manipulating a graphical user interface and interactive input system employing the same |
US20130078990A1 (en) * | 2011-09-22 | 2013-03-28 | Mikyung Kim | Mobile device and method for controlling reproduction of contents in mobile device |
US20130201123A1 (en) * | 2012-02-06 | 2013-08-08 | Lg Electronics Inc. | Mobile terminal and electronic communication method using the same |
US20140009394A1 (en) * | 2012-07-06 | 2014-01-09 | Lg Electronics Inc. | Mobile terminal, image display device and user interface provision method using the same |
US20140033058A1 (en) * | 2012-07-30 | 2014-01-30 | Plantronics, Inc. | Use of multiple mobile devices to control an application |
US20140164966A1 (en) * | 2012-12-06 | 2014-06-12 | Samsung Electronics Co., Ltd. | Display device and method of controlling the same |
US20140365923A1 (en) * | 2013-06-10 | 2014-12-11 | Samsung Electronics Co., Ltd. | Home screen sharing apparatus and method thereof |
Cited By (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140282103A1 (en) * | 2013-03-16 | 2014-09-18 | Jerry Alan Crandall | Data sharing |
US20160110075A1 (en) * | 2013-03-16 | 2016-04-21 | Jerry Alan Crandall | Data Sharing |
US20160110153A1 (en) * | 2013-03-16 | 2016-04-21 | Jerry Alan Crandall | Data Sharing |
US20160110074A1 (en) * | 2013-03-16 | 2016-04-21 | Jerry Alan Crandall | Data Sharing |
US20160110072A1 (en) * | 2013-03-16 | 2016-04-21 | Jerry Alan Crandall | Data Sharing |
US20160110073A1 (en) * | 2013-03-16 | 2016-04-21 | Jerry Alan Crandall | Data Sharing |
US9563341B2 (en) * | 2013-03-16 | 2017-02-07 | Jerry Alan Crandall | Data sharing |
US9645720B2 (en) * | 2013-03-16 | 2017-05-09 | Jerry Alan Crandall | Data sharing |
US20200084116A1 (en) * | 2013-11-01 | 2020-03-12 | Google Llc | User-configurable delivery of informational events |
WO2016003943A1 (en) * | 2014-06-29 | 2016-01-07 | Tetherball Technology, LLC | Automatic wireless grid access sharing and monetization from a mobile device app |
US9936528B2 (en) | 2014-06-29 | 2018-04-03 | Tetherball Technology, LLC | Automatic wireless grid access sharing and monetization from a mobile device app |
Also Published As
Publication number | Publication date |
---|---|
US20120191832A1 (en) | 2012-07-26 |
US8554897B2 (en) | 2013-10-08 |
WO2012102416A1 (en) | 2012-08-02 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8554897B2 (en) | Data sharing between smart devices | |
JP5819965B2 (en) | System and apparatus | |
US20220368742A1 (en) | Shared-content session user interfaces | |
US20220365740A1 (en) | Shared-content session user interfaces | |
KR101869485B1 (en) | Radar-based gesture-recognition through a wearable device | |
CA2823542C (en) | Methods and systems for displaying content on multiple networked devices with a simple command | |
WO2022121358A1 (en) | Information display method and apparatus | |
US9406103B1 (en) | Inline message alert | |
US20110285658A1 (en) | Information processing device, information processing method, and program | |
US20120102438A1 (en) | Display system and method of displaying based on device interactions | |
US20140006949A1 (en) | Enhanced user interface to transfer media content | |
US20150143423A1 (en) | Apparatus, method, and system for controlling device based on user interface that reflects user's intention | |
CN112073767B (en) | Display equipment | |
EP3243333A1 (en) | Automatic home screen determination based on display device | |
US11627360B2 (en) | Methods, systems, and media for object grouping and manipulation in immersive environments | |
WO2020073334A1 (en) | Extended content display method, apparatus and system, and storage medium | |
WO2023050737A1 (en) | Resource presentation method based on live streaming room, and terminal | |
US9495064B2 (en) | Information processing method and electronic device | |
US20210326010A1 (en) | Methods, systems, and media for navigating user interfaces | |
CN107172472B (en) | Running touchscreen applications on touchless capable display devices using remote controls | |
EP3386204A1 (en) | Device and method for managing remotely displayed contents by augmented reality | |
US20150082347A1 (en) | Delivering Video Program by Digital Television System and Mobile Communication Device | |
TW201209704A (en) | Image-based control method, processing method and system thereof | |
US9875012B2 (en) | Media sharing between devices using drag and drop gesture | |
KR102324624B1 (en) | Methods, systems and media for presenting media content previews |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |