US20070171192A1 - Screen image presentation apparatus and method for mobile phone - Google Patents

Screen image presentation apparatus and method for mobile phone Download PDF

Info

Publication number
US20070171192A1
US20070171192A1 US11/633,624 US63362406A US2007171192A1 US 20070171192 A1 US20070171192 A1 US 20070171192A1 US 63362406 A US63362406 A US 63362406A US 2007171192 A1 US2007171192 A1 US 2007171192A1
Authority
US
United States
Prior art keywords
character
item
house
items
character house
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US11/633,624
Inventor
Jeong Seo
Jong Lee
Wei Park
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Samsung Electronics Co Ltd
Original Assignee
Samsung Electronics Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Samsung Electronics Co Ltd filed Critical Samsung Electronics Co Ltd
Assigned to SAMSUNG ELECTRONICS CO., LTD. reassignment SAMSUNG ELECTRONICS CO., LTD. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: LEE, JONG KERL, PARK, WEI JIN, SEO, JEONG WOOK
Publication of US20070171192A1 publication Critical patent/US20070171192A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/72427User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality for supporting games or graphical animations
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04BTRANSMISSION
    • H04B1/00Details of transmission systems, not covered by a single one of groups H04B3/00 - H04B13/00; Details of transmission systems not characterised by the medium used for transmission
    • H04B1/38Transceivers, i.e. devices in which transmitter and receiver form a structural unit and in which at least one part is used for functions of transmitting and receiving
    • H04B1/40Circuits
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/7243User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72448User interfaces specially adapted for cordless or mobile telephones with means for adapting the functionality of the device according to specific conditions
    • H04M1/72451User interfaces specially adapted for cordless or mobile telephones with means for adapting the functionality of the device according to specific conditions according to schedules, e.g. using calendar applications
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M2250/00Details of telephonic subscriber devices
    • H04M2250/56Details of telephonic subscriber devices including a user help function
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W52/00Power management, e.g. TPC [Transmission Power Control], power saving or power classes
    • H04W52/02Power saving arrangements
    • H04W52/0209Power saving arrangements in terminal devices
    • H04W52/0261Power saving arrangements in terminal devices managing power supply demand, e.g. depending on battery level
    • H04W52/0267Power saving arrangements in terminal devices managing power supply demand, e.g. depending on battery level by controlling user interface components
    • H04W52/027Power saving arrangements in terminal devices managing power supply demand, e.g. depending on battery level by controlling user interface components by controlling a display operation or backlight unit
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D30/00Reducing energy consumption in communication networks
    • Y02D30/70Reducing energy consumption in communication networks in wireless communication networks

Definitions

  • the present invention relates to portable phone. More particularly, the present invention relates to an idle mode screen image presentation apparatus and method for a mobile phone.
  • a portable terminal is typically a cordless electronic device such as a mobile phone.
  • Recent mobile phones are implemented with various functions in addition to basic voice communication.
  • Such phones include a camera phone equipped with a camera module, a digital TV phone equipped with a Digital Multimedia Broadcast (DMB) or Digital Video Broadcast (DVB) module, and an MP3 phone equipped with an MPEG-1 Audio Layer 3 (MP3) module. That is, the mobile phone has developed into a multipurpose electronic device.
  • DMB Digital Multimedia Broadcast
  • DVD Digital Video Broadcast
  • MP3 MPEG-1 Audio Layer 3
  • an advanced mobile phone with an advanced multimedia processing ability is provided with a liquid crystal display (LCD) to display multimedia data.
  • LCD liquid crystal display
  • Various screen display techniques have been researched and developed to effectively display information. That is, a screen display scheme is designed to support viewing multimedia data while simultaneously displaying information related to a basic task. For example, a mobile phone can be implemented to display a preset image while the mobile phone is in an idle state. The image of the idle screen view is downloaded from a website or taken by a built in camera.
  • conventional mobile phones limit designs to screen images to be presented on the display by themselves. Although there has been a screen image design method in which the screen image is designed with a background image, text, and clock selected by a user, most of the conventional screen image design techniques do not facilitate the design of a dynamic screen image with various character images.
  • An aspect of exemplary embodiments of the present invention is to address at least the above problems and/or disadvantages and to provide at least the advantages described below. Accordingly, an aspect of exemplary embodiments of the present invention is to provide a screen image presentation apparatus and method for a mobile phone.
  • a screen image presentation apparatus and method for a mobile phone are provided. This facilitates the presentation of a character house with an artificial intelligent character, items, and background as a screen image.
  • the intelligent character reacts and adapts to a change in the conditions of the mobile phone in the form of an animation, sound, and text.
  • a screen image presentation apparatus and method for a mobile phone is provided.
  • This screen image presentation apparatus and method is capable of providing a character house as an animated screen image for a specific operation mode of a mobile phone.
  • the character house includes a character, background, function items, decoration items and motion items.
  • the character illustrates conditions and emotions reflecting states of the mobile phone in the form of an animation, sound, and text.
  • the function items illustrate system information on the mobile phone. Decoration items improve aesthetics of the screen image and motion items move along respective movement lines in the screen house.
  • a screen image presentation apparatus and method for a mobile phone are provided.
  • This apparatus and method facilitates a user's ability to design a character house, by selecting a character, background, and items, to be used as a screen image in an idle mode of the mobile phone.
  • the character house is used as a screen image for a specific operation mode of the mobile phone.
  • the character house consists of a background, character, and items and is used as a screen image for a specific operation mode of the mobile terminal.
  • a screen image presentation apparatus and method for a mobile phone facilitates a user's ability to select a partitioned sectors of a character house to see the selected sector in an enlarged form such that the user can perform a fine adjustment of a position of an item located in the selected sector.
  • the character house consists of a background, character, and items, and is used for a screen image for a specific operation mode of the mobile terminal.
  • a screen image presentation apparatus of a mobile phone includes a memory, a controller and a display.
  • the memory stores a character house including a character for presenting a condition of the mobile phone.
  • the controller controls to present the character house at a predetermined presentation time and the display displays the character house under the control of the controller.
  • a screen image presentation method for a mobile phone According to the screen image presentation method, configuration information on a character house is verified with a character that represents a condition which is determined by usage pattern of the mobile phone. The character house is presented in a specific operation state of the mobile phone on the basis of the configuration information.
  • a screen image presentation method for a mobile phone configuration information of a character house is registered in a program mode.
  • the character house with a background and items is edited in an editing mode.
  • the configuration information of the character house is checked and the character house is presented with a character, items, and background on the basis of the configuration information of the character house.
  • FIG. 1 is a block diagram illustrating a configuration of a mobile phone employing a screen image presentation apparatus and method according to an exemplary embodiment of the present invention
  • FIGS. 2 a, 2 b and 2 c are views illustrating a concept of a character house of a screen image presentation apparatus according to an exemplary embodiment of the present invention
  • FIG. 3 is a screen image illustrating an exemplary character house according to an exemplary embodiment of the present invention.
  • FIGS. 4 a to 4 n are screen images illustrating steps for implementing exemplary character animations in a screen image presentation technique according to an exemplary embodiment of the present invention
  • FIGS. 5 a to 5 c are screen images illustrating steps for implementing an animated background in a screen image presentation technique according to an exemplary embodiment of the present invention
  • FIG. 6 illustrates screen images that illustrate background images for implementing an animated background in a screen image presentation technique according to an exemplary embodiment of the present invention
  • FIG. 7 a is screen image illustrating a character quotient module on the idle mode screen in a screen image presentation technique according to an exemplary embodiment of the present invention
  • FIG. 7 b is an enlarged view illustrating the character quotient module of FIG. 7 a;
  • FIG. 8 is a drawing illustrating a table of emoticons to be displayed with the quotient module of FIG. 7 b.
  • FIGS. 9 a to 9 l are views illustrating the presentation of function items in a screen image presentation technique according to an exemplary embodiment of the present invention.
  • FIGS. 10 a to 10 c are views illustrating steps of presenting motion items in a screen image presentation technique according to an exemplary embodiment of the present invention
  • FIGS. 11 a to 11 c are views illustrating steps of presenting decoration items in a screen image presentation technique according to an exemplary embodiment of the present invention
  • FIG. 12 illustrates screen images illustrating character behaviors in the screen image presentation technique according to an exemplary embodiment of the present invention
  • FIG. 13 is a screen image illustrating a character animation with an event in a screen image presentation technique according to an exemplary embodiment of the present invention
  • FIG. 14 is a flowchart illustrating a character house presentation procedure of a screen image presentation method according to an exemplary embodiment of the present invention.
  • FIG. 15 is a flowchart illustrating the background presentation procedure of FIG. 14 ;
  • FIG. 16 is a flowchart illustrating the item presentation procedure of FIG. 14 ;
  • FIG. 17 is a flowchart illustrating a procedure for presenting a motion item in the screen image presentation technique according to an exemplary embodiment of the present invention.
  • FIGS. 18 a to 18 d are flowcharts illustrating a procedure for presenting function items in the screen image presentation technique according to an exemplary embodiment of the present invention
  • FIG. 19 is a flowchart illustrating a character presentation procedure of the screen image presentation method according to an exemplary embodiment of the present invention.
  • FIG. 20 is a flowchart illustrating a greeting behavior presentation procedure of FIG. 19 ;
  • FIG. 21 is a flowchart illustrating a greeting event presentation procedure that occurred during a greeting behavior of the character of the screen image presentation technique according to an exemplary embodiment of the present invention
  • FIGS. 22 a and 22 b are flowcharts illustrating an animation behavior presentation procedure of the screen image presentation technique according to an exemplary embodiment of the present invention
  • FIGS. 23 a and 23 b illustrate the character shift information determination procedure in detail
  • FIG. 24 is a flowchart illustrating a screen house presentation procedure of a screen image presentation method for a mobile phone according to another exemplary embodiment of the present invention.
  • FIG. 25 is a state diagram illustrating a character house design procedure of a screen image presentation method according to an exemplary embodiment of the present invention.
  • FIG. 26 is a state diagram illustrating transitions between the character house configuration mode and the character house editing mode of FIG. 25 ;
  • FIG. 27 is a screen image illustrating a character house configuration screen in the character house configuration mode of FIG. 25 ;
  • FIG. 28 is a state diagram illustrating transitions between the character house configuration mode and the character house editing mode of FIG. 25 ;
  • FIG. 29 illustrates screen images illustrating a procedure for setting a “character house presentation on/off” option on the character house configuration screen of FIG. 25 ;
  • FIG. 30 is a state diagram illustrating transitions among the character house editing mode and specific item editing modes of FIG. 25 ;
  • FIG. 31 illustrates a screen image illustrating an example character house editing screen in the character house editing mode of FIG. 25 ;
  • FIG. 32 is a state diagram illustrating transitions between the character house editing mode and the background selection mode of FIG. 25 ;
  • FIG. 33 illustrates screen images illustrating a procedure for selecting a background on the background selection mode of FIG. 32 ;
  • FIG. 34 is a state diagram illustrating transitions between the character house editing mode and item editing modes of FIG. 25 ;
  • FIGS. 35 a to 35 d illustrate screen images illustrating a procedure for adding an item to the character house in the character house editing mode of FIG. 34 ;
  • FIG. 36 is a state diagram illustrating transitions between the character house editing mode and the selected item editing mode via the edited item shortcut selection mode of FIG. 25 ;
  • FIGS. 37 a to 37 c illustrate screen images illustrating a procedure for assigning shortcut key in the edited item shortcut selection mode of FIG. 36 .
  • a “screen image” is an image presented on a display screen of the mobile phone.
  • “House,” “avatar house,” and “character house” are used interchangeably to refer to a visible object presented on screen in a specific mode of the mobile phone. The house is displayed with characters, items, and background.
  • “Item” is a visible object for enriching the house and includes decoration items, function items for notifying states of specific functions, and motion items that move along preset paths.
  • Background is an image forming a background graphics of the house.
  • “Character” is a visible object that resides in the character house and can express a state of the mobile phone in an emotional manner.
  • the character includes fictional characters represented by a human, avatar, or animal, among others that can express various emotions. Accordingly, the character can express emotions.
  • the character house can be correlated with a background, a character, and at least one item.
  • FIG. 1 is a block diagram illustrating a configuration of a mobile phone employing a screen image presentation apparatus and method according to an exemplary embodiment of the present invention.
  • the mobile phone includes a controller 110 , data processing unit 120 , radio frequency (RF) unit 123 , audio processing unit 125 , voice synthesizer 127 , memory 130 , keypad unit 141 , and display unit 143 .
  • RF radio frequency
  • the RF unit 123 facilitates communication between the mobile phone and other terminals over a wireless channel.
  • the RF unit 123 includes an RF transmitter for up-converting and amplifying the frequency of a transmitted signal, and an RF receiver for low-noise amplifying a received signal and down-converting the frequency of the received signal.
  • the data processing unit 120 includes a transmitter for coding and modulating the transmitted signal, and a receiver for demodulating and decoding the received signal.
  • the data processing unit 120 may include a modem and a codec.
  • the voice synthesizer 129 performs a function of converting a text message output from the controller 110 or received from outside into a voice signal.
  • the keypad unit 141 is provided with a plurality of alphanumeric keys for input data and functions key for setting various functions.
  • the memory 130 includes a program memory and a data memory.
  • the program memory stores programs for processing incoming and outgoing calls, and programs for presenting the character house for a specific state of the mobile phone and controlling reactions of the characters in response to events caused by operations of the mobile phone and for structuring the screen image comprising the character house with the characters, according to an exemplary embodiment of the present invention.
  • the data memory includes a character memory for storing character images and information on the characters for performing the exemplary embodiment of the present invention and a text memory for storing text data for emotional expressions of the characters.
  • the data memory also temporarily stores data generated while the programs are executed.
  • character house image is interchangeably used to refer to the screen image presented in a specific mode of the mobile phone.
  • the specific mode or specific state may refer to the idle mode or idle state.
  • the character memory may include a character storage, an item storage and a background storage.
  • the character storage stores information on the structures of the characters and images thereof.
  • the item storage stores information on the structures of the items and image thereof.
  • the background storage stores background images.
  • the character house structure may be stored in the form of information on a combination of the character, item, and background image, and the images stored in the respective storages may be used individually.
  • a character is provided with information on character animations expressing emotional states for the mobile phone, character images for expressing emotions in response to text messages (for example, SMS, EMS, and MMS, among others), and character images for expressing emotions in response to events occurring in the mobile phone.
  • text messages for example, SMS, EMS, and MMS, among others
  • character images for expressing emotions in response to events occurring in the mobile phone are assumed.
  • the memory for storing the images operates in a file structure.
  • a character file structure includes a header and character images, and the header contains information on identifiers, start delimiters, a number of images, display positions and names of animations.
  • the headers and character images are transmitted after being processed in the form of a transmission file.
  • the character house is displayed on the screen together with a background image, a character, and at least one item in a specific mode (for example, an idle mode) of the mobile phone. That is, the character house is designed by arranging the items on the background image for a specific mode screen (for example, an idle mode screen), the items and the background images being selected by the user.
  • the file structure of the character house may include a header and a plurality of images, and the header may include a header of the character house and the item headers of the items associated with the character house.
  • the character house header contains information on the character house identifier, the positions of the characters on the screen, the sizes of the characters, and the names of the characters; and each item header contains information on the item index, the position of the item in relation with the character house, and the name of the item.
  • a character house transmission mode the headers of the character house and the items associated with the character house and the background and item images are transmitted in the form of a compressed file.
  • the items means the images presented in association with the character house on the screen. Each item may be displayed in a specific association with the character house or regardless of the character house. If the item is not associated with the character house, the item should be transmitted individually.
  • the item includes a header and an image, and the item header contains information on the item index, the position of the item in relation to the character house, and the name of the item. According to an exemplary implementation, if the item is not associated with the character house, the information on the position of the item has a default value.
  • the character is used for presenting the information on the states and event occurrences of the mobile phone.
  • the character house is used to indicate the specific operation mode (for example, the idle mode) of the mobile phone.
  • the item is used to show the functions of the mobile phone, decorate the character house, and show the screen dynamic.
  • the background image is used to provide a background of the characters and items associated with the character house.
  • the controller 110 controls general operations of the mobile phone.
  • the controller 110 may be integrated with the data processing unit 120 .
  • the controller 110 , data processing module 120 , audio processing unit 125 , and voice synthesizer 127 can be integrated into a mobile station modem (MSM) chip.
  • MSM mobile station modem
  • the controller 110 controls to select the characters, items, background image, character house, texts, and sounds.
  • the controller arranges the selected images and texts at the present positions on the screen of the display unit 143 and outputs the texts as voice sounds under the support of the voice synthesizer 127 .
  • the display unit 143 displays the state of the mobile phone under the control of the controller 110 .
  • the display unit 143 may be implemented with a liquid crystal display (LCD) together with a LCD controller, extra memory for storing the display data, and LCD display device. If the display unit 143 is implemented with a touchscreen function, the display unit 143 can operate an input device.
  • LCD liquid crystal display
  • the characters represent a graphical user interface (GUI) including the character displayed on the screen of the display unit 143 in the specific mode of the mobile phone.
  • GUI graphical user interface
  • the character house is a space in which the character with an artificial intelligence resides such that the user can experience a virtual space with a virtual talking character. The user can quickly learn the functions of the mobile phone with the assistances of the character house and enjoy with the vivid characters.
  • the character house has a function that expresses a dynamic character.
  • the character house can express the condition and emotion of the current character in the animation, sound, and text according to a character scenario associated with the character.
  • the character house has a function to present a background of the screen for a specific mode (especially, the idle mode) of the mobile phone that is changed in accordance with a progress of time.
  • the character house has a function for displaying items.
  • the items displayed in association with the character house may be classified by function. In this exemplary embodiment of the present invention, the items can be classified into three categories.
  • the first is the function item for presenting the functions and system information on the mobile phone in order to enhance the user's convenience.
  • the second is the decoration item for increasing aesthetic satisfaction by adding various images to the character house.
  • the third is the motion items that make the character house screen dynamic by adding the motion images (for example, the cloud or bird, among others).
  • the character house may further provide a greeting function with which the character associated with the character house gives greetings with appropriate sounds at preset times.
  • An additional function such as allowing a user to give a present to the character may also be implemented.
  • FIGS. 2 a to 2 c are views illustrating a concept of a character house of a screen image presentation apparatus according to an exemplary embodiment of the present invention.
  • FIG. 2 a is an illustration of a house setting screen for selecting a character house to be displayed on the screen.
  • the house setting screen is displayed on the screen under the control of a character setting program. If the character setting program is executed, 4 character house images are displayed in corresponding tiled windows on the screen of the display unit 143 . When a number of the character house images are less than 4, the windows that are not allocated for the character house image will be empty.
  • the controller 110 operates a character engine which is structured as illustrated in FIG. 2 b.
  • the character engine stores information of the selected character house into a character house information storage module 210 and executes a character house expression module 220 upon activation. That is, the character engine includes the character house information storage module 210 and the character house expression module 220 .
  • the character house information storage module 210 is a module for storing information on the character houses and the character to be displayed in association with each other. According to an exemplary embodiment of the present invention, the character house information storage module 210 includes character house information storage 213 , system information storage 215 , and character information storage 217 .
  • the character information storage 217 stores emotion information, status information, quotient information, and text and sound information associated with the character's emotions and statuses.
  • the character information storage 217 also stores animation information for animating the character in a specific status.
  • the mobile phone processes various events which have occurred while operating.
  • the mobile phone also updates the quotients of the characters that correspond to the events.
  • the quotient information includes a physical quotient (HQ), emotion quotient (EQ), intelligent quotient (IQ), and social quotient (SQ), that are used for determining the status and emotion of the character.
  • the controller 110 accumulates the quotients associated with the events whenever an event is processed and determines the status of the mobile phone on the basis of the quotient analysis results.
  • the emotion of the character is determined according to the quotients and user's feedback analysis results.
  • the emotion of the character can be determined in accordance with a content related to the event (for example, a text message) that is obtained during the process.
  • the text and sound information associated with the emotion can be stored together with the character expression.
  • the character information storage 217 includes information on the animation of the character that is expressed in a specific mode (for example, an idle mode) of the mobile phone.
  • the system information storage 215 stores system information on the function items for displaying a memory capacity of the mobile phone, scheduled alarms and unread mail.
  • the items associated with the character house includes the decoration items, motion items, and function items.
  • the function items present the operations and functions of the mobile phone. According to an exemplary embodiment of the present invention, the items for presenting a memory status, preset alarm times, and unread mail are described as examples of the function items.
  • the character house information storage 213 stores the information on the character house that is defined by a character house setting program such that the character house information includes the types and positions of the items and the type of the background associated with the character house.
  • the character house information storage 213 stores the information on the character house images except for the character and the function items.
  • the character house information 213 stores the information on the decoration items, motion items and the background images.
  • the character house expression module 220 includes a character display module 223 , an item display module 225 and a background display module 227 .
  • the character display module 223 renders the character animation of the character house.
  • the character display includes a mobility and entrance of the character and a text balloon or a sound effect associated with the behavior of the character.
  • the character display module 223 can make the character express special motions such as a greeting.
  • the item display module 225 displays the items at preset positions with reference to the character house information on the current character house and the system information. The items are displayed in a different manner by item category (decoration, motion, and function).
  • the background display module 227 displays the background image in accordance with the information stored in the character house information storage 213 .
  • the background image can be configured such that the brightness of the background image changes over time.
  • FIG. 2 c illustrates an exemplary character house screen displayed by the character engine of FIG. 2 b.
  • the character house is displayed in a specific mode (for example, an idle mode) of the mobile phone and includes a character, items and a background.
  • FIG. 3 is a screen image illustrating an exemplary character house according to an exemplary embodiment of the present invention.
  • the character house screen provides information on the detailed status, which is not provided with a normal idle mode screen, for the mobile phone together with decoration effects, as illustrated in FIG. 3 .
  • a character 230 is displayed in the form of an animated character in association with the character house.
  • the character 230 behaves in a motion preset for the idle mode of the mobile phone to alarm at preset times in corresponding animation together with a text and sound effect.
  • the character 230 can be set to animate for giving items as presents for the character house in a random manner.
  • the character 230 expresses the emotion in the animation according to the state of the mobile phone and an emotion model together with the sound and text.
  • the background 240 is represented by a background image of the character house and the brightness of the background can be changed over time.
  • the items 250 give effects for decorating the character house and inform the operations of the mobile phone.
  • the function items 251 , 253 , 255 , and 259 decorate the character house and present the statuses of the preset functions and operations.
  • the function item 251 is a clock item that illustrates a preset alarm time
  • the function item 253 is a mailbox item for illustrating the amount of unread mail. If multiple alarms are set, the closest alarm time is displayed.
  • the function item 255 is a memory monitor item for displaying the state of the memory.
  • the decoration item 259 is an item for decorating the character house.
  • the motion item (not shown) can be activated for decorating the character house.
  • the character quotient module 260 includes a character quotient, a physical quotient, an intelligent quotient (IQ) and an emotional quotient (EQ).
  • the social quotient (SQ) is represented by values assigned for the events related to the communication.
  • the physical quotient (HQ) is represented by values assigned for events related to the performance for the mobile phone.
  • the intelligent quotient (IQ) is represented by values assigned for an amount of information stored in the mobile phone and the emotional quotient (EQ) is represented by values assigned for events related to the multimedia.
  • the character quotient module 260 can be configured to be fixed at a position or floats across the screen.
  • the character 230 of the character house has the following three basic expression factors: animation, text and speech.
  • Animation facilitates the character's 230 ability to express a state and an emotion while moving.
  • a text facilitates the character's 230 expression of a state and emotion in text.
  • the text can be presented in the form of a text balloon.
  • Speech facilitates the character's 230 expression of a state and emotion through voice or sound effects.
  • the character 230 can express various states and emotions, and the state and emotion of the character 230 are determined in accordance with a state decision model algorithm and an emotion mode decision algorithm as a character scenario dependent on the variation of the quotient values of the character.
  • the decision on the state and emotion mode of the character on the basis of the character scenario is disclosed in the Korean Patent Application No. 2005-107174.
  • the character 230 of the character house expresses the state information and the emotion mode using the character image together with a unique voice effect and text balloon.
  • the character 230 can be implemented with an animation in which the character 230 moves around the character house and disappears outside the screen and then appears from outside. That is, the character 230 expresses the state and emotion with a motion, sound effect, and text, if the time to express the state and emotion has arrived. The state and emotion of the character 230 is expressed when the quotients reach respective threshold values, rather than always. However, the character house is always displayed in the specific mode screen. For example, if the mobile phone enters an idle mode, the character house is displayed on the screen. The character house screen disappears when a preset display time expires or when the mobile phone enters another operation mode. If the character house is displayed at the idle mode, the character 230 expresses the state and emotion in motion and with other animations.
  • FIGS. 4 a to 4 h are screen images illustrating steps for implementing an exemplary character animation in a screen image presentation technique according to an exemplary embodiment of the present invention.
  • FIGS. 4 i to 4 n are screen images illustrating steps of another exemplary character animation in a screen image presentation technique according to an exemplary embodiment of the present invention.
  • the character 230 moves in a preset behavior pattern in the character house.
  • the character 230 appears at a random position on the screen, makes a speech expression and then moves around. If the character house disappears on the screen while the character moves around the character house, the character 230 appears at a random position and repeats the behavior after a predetermined time has elapsed.
  • the motion, sound effect, and text can be selectively disabled.
  • the character house can be expressed through six steps as followings.
  • a character 230 appears at a position in the character house.
  • the character 230 can appear with a fade-in effect. That is, the character 230 can be implemented with an alpha blending technique so that the transparency of the character 230 changes from 100% to 0% gradually. For example, the transparency reduces by 25% per 0.5 second such that the character 230 is gradually faded in to completely appear in 2 seconds.
  • the character 230 starts appearing at a random position.
  • the character 230 is preferably implemented so that the character is at a position in which the character does not completely appear. That is, the character 230 is configured to appear at a random position but where the character 230 can be completely shown across the border of the screen.
  • a sound effect is output in accordance with the state of the character and with a notification icon, as illustrated in FIG. 4 c.
  • the sound effect is output at once and repeated one more time when the character reappears.
  • the sound effect may also be cancelled.
  • the character 230 is animated as illustrated in FIG. 4 d.
  • the character animation process is repeated until the character 230 disappears.
  • the character animation is performed at a fixed position.
  • the state and emotion of the character 230 is expressed with a text after the character animation starts.
  • the text is expressed in the text balloon in consideration of the time taken for the user to read the text.
  • the text balloon is displayed in the form of a semitransparent image with about 30% of transparency and is open in about 3 seconds in consideration of the reading time.
  • the character 230 moves as illustrated in FIGS. 4 f and 4 g.
  • the movement direction of the character 230 can be changed according to the position at which the character 230 appears. That is, the position at which the character 230 appears is verified such that the character 230 moves in a direction to which the character can secure the longest distance for movement.
  • the character 230 moves at a predetermined speed.
  • the character 230 is preferably set to move at the speed of about 20 pixels per second.
  • the character 230 disappears across the border of the screen. If the character 230 disappears completely, an initial idle mode screen is displayed as illustrated in FIG. 4 h identical with FIG. 4 a. After a predetermined time (about 2 seconds) elapses, steps 4 a to 4 g are repeated.
  • the character animation illustrated in FIGS. 4 i to 4 h is an example in which the character does not move.
  • the character is displayed in the idle mode screen.
  • the preset items and background images are displayed without the character.
  • FIG. 4 j the character appears on the screen.
  • the character appears at a predetermined position on the screen.
  • the appearance position is changed to the bottom right position in relation to the center of the screen.
  • the appearance position can be changed according to the size of the character.
  • the character can be set to appear with a fade-in effect in such a manner that the transparency of the character changes from 100% to 0% gradually. For example, the transparency reduces by 25% per 0.5 second such that the character is gradually faded in to completely appear in 2 seconds.
  • a sound effect is output in accordance with the state of the character and with a notification icon as illustrated in FIG. 4 k.
  • the sound effect is output at once and repeated one more time when the character reappears.
  • a text is displayed to express the current state and emotion of the character.
  • the text is presented in the text balloon with a transparency of 30%.
  • the text can be displayed in synchronization with a sound.
  • the text balloon is expressed in consideration of the time taken for the user to read the text. For example, the text balloon can be set to disappear after 1.5 seconds.
  • the character is animated as illustrated in FIG. 4 m.
  • the character animation process is repeated until the character disappears in the character house as illustrated in FIG. 4 n.
  • the character animation process is performed in a certain order so that the items 250 and background image 240 are displayed, the character 230 appears in a fade-in effect, the sound effect is output for notifying the state of the mobile phone, the text is presented for notifying the state of the mobile phone, and the character animation is performed in the character house.
  • the character animation is a motion behavior of the character without moving around the screen, and the character movement is a behavior of the character moving around the screen.
  • the character animation and the movement may be performed simultaneously.
  • the character 230 may also appear simultaneously with the items 250 and the background.
  • the character 230 may appear in the form of an opaque object, such as with a 0% transparency. The output of sound effect and text may be skipped.
  • the presentation of the background can be performed so that the background image changes according to the progress of the time.
  • a fixed background image can be used.
  • FIGS. 5 a to 5 c are screen images illustrating steps for implementing an animated background in a screen image presentation technique according to an exemplary embodiment of the present invention.
  • FIG. 6 illustrates screen images illustrating background images for implementing an animated background in a screen image presentation technique according to an exemplary embodiment of the present invention.
  • the background images of FIG. 6 can be used for the animated background or the fixed background.
  • the animated background is implemented with several background images that are presented in series over a period of time.
  • a day is divided into time periods, such as morning, afternoon, evening, and night.
  • the background is placed with different background images for the respective time periods.
  • the background images of the character house change according the specific time period if an animation background 240 is set by the user. Since the change of the background is performed in accordance with the progress of the time, the background of the character house that changed with different brightness background images is assigned for the morning, day, and night as illustrated in FIGS. 5 a to 5 c.
  • the time periods are defined as illustrated in table 1.
  • the character can be set to present corresponding greetings at the beginning of each time period.
  • the background may be set on the character house setting screen.
  • FIG. 6 illustrates the background images that may be selected for the background.
  • the background images may be used for the animated background in which background images are changed in accordance with the time period or may be used for a fixed background.
  • the background images can be used for the animated background of the character house or can be used for the fixed background of the character house.
  • Each background image can be used for the animated background by only changing the brightness of the background image for expressing the time periods.
  • the character house can present information on items that illustrate the system information and on the quotients of the character, as well as the character 230 and the background 240 .
  • FIG. 7 a is screen image illustrating a character quotient module on the idle mode screen in a screen image presentation technique according to an exemplary embodiment of the present invention.
  • FIG. 7 b is an enlarged view illustrating the character quotient module of FIG. 7 a.
  • the character quotient module 260 notifies the current status of the character 230 .
  • the character quotient module 260 illustrates the state and emotion of the character 230 with the HQ, EQ, IQ, and SQ status gauges. Since the state and emotion of the character are presented with the quotient status gauges, the character can attract the user's interest.
  • the character quotient module 260 can be presented at a preset position in the character house. In this exemplary embodiment of the present invention, the character quotient module is presented at the top left area of the screen, as illustrated in FIG. 7 a.
  • the character quotient module 260 can be set to be visible or invisible according to a user selection.
  • the character quotient module 260 includes a quotient information display region 267 , an emotion icon display region 263 , and a status display region 265 .
  • the emotion icon display region 263 can be implemented to display emoticons representing the emotions of the character.
  • the status display region 265 presents the status of the character in text and the quotient information display region 267 displays the levels of the quotients with color bars (gauge).
  • the emotions that can be presented in the emotion icon display region 263 includes “happy,” “normal,” “angry,” “sad,” and “tired,” and the emotions are linked to corresponding images in the form of a table as illustrated in FIG. 8 .
  • the emotion of the character can be presented by the emoticon of the character quotient module corresponding to the current emotion of the character in addition with the character animations.
  • the status information of the character can be presented at the status display region 265 in the form of a text.
  • Each character can have various statuses that can be expressed in the form of a text.
  • the character quotient information is displayed in the form of gauges in the quotient information display region 267 .
  • four quotients such as, HQ, EQ, IQ, and SQ are measured and presented in the quotient information display region 267 .
  • the quotients are presented by means of colored gauges, such as a red gauge for SQ, a green gauge for HQ, a purple gauge for IQ, a blue gauge for EQ, as shown in FIG. 7 b.
  • Each gauge presents the level of the quotient in a percentage (%) without a text based notification.
  • the emotion icon display region 263 and a status display region 265 can be deleted.
  • the items and expressions such as the greetings can be bought in an item shop on the web or received as a gift.
  • the items that can be presented in the character house include the function items, decoration items, and motion items.
  • the positions of the items on the screen can be set in the character house setting screen. Presentation of the items is explained based on an order of the function item, motion item, and decoration item.
  • the function items are items that provide specific system information of the mobile phone.
  • the function items include a memory monitor item, an alarm item, and a mailbox item. Additional function items may also be added when required.
  • the mailbox item can be a message item, which is used to explain the function item.
  • the function items can be implemented for a signal function or can be an item grouping the function items comprising similar functions.
  • the mailbox item can present a variety of the messages.
  • the function item integrates a short message service (SMS) item, an email message service (EMS) item, and a multimedia message service (MMS) item.
  • SMS short message service
  • EMS email message service
  • MMS multimedia message service
  • the controller 110 can display the function items integrated in the multifunction item such that the user can select one of the items.
  • FIGS. 9 a to 9 l are views illustrating how to present function items in a screen image presentation technique according to an exemplary embodiment of the present invention.
  • the functional item can be displayed in a normal mode (see FIG. 9 a ), a selection mode (see FIG. 9 b ), and an enable mode (see FIG. 9 c ). Presentation of the function item with a single function is described with reference to FIGS. 9 a to 9 c.
  • the function items are arranged at the preset positions on the screen.
  • the mailbox item, the clock item, and the memory monitor item are examples displayed on the screen.
  • the function items are displayed as basic item images. If an event occurs while the function items are displayed in the normal mode, the basic item image of the function item related to the event is animated.
  • the mobile phone enters a selection mode for displaying the function item information if an item display mode key (a special key such as “clear key” provided to the mobile phone) is input.
  • an item display mode key a special key such as “clear key” provided to the mobile phone
  • the function items are displayed with a semitransparent layer in the selection mode, and the selected item is zoomed in to be displayed in an enlarged form.
  • other function items can be selected by keypad manipulation. If an OK key is input while an item is selected, the item is activated in the enable mode.
  • the item image of the selected function item is displayed together with the information on the selected function item.
  • the up/down direction keys provided in the keypad unit, it is possible to navigate to other applets or return to the normal mode shown in FIG. 9 a.
  • Left/right direction keys allow navigating to and activating other function items.
  • FIGS. 9 d to 9 f A structure and presentation of a function item are described with the memory monitor item as an example hereinafter with reference to FIGS. 9 d to 9 f.
  • the memory monitor item is a function item for display memory utilization information.
  • FIG. 9 d illustrates a memory monitor item image in the normal mode. If the memory usage is over 50% in the normal mode, the memory monitor item is animated as illustrated in FIG. 9 e. If the memory monitor item is selected by the keypad manipulation, the memory monitor item is displayed on the screen in the form of an enlarged memory monitor item image presenting the memory utilization with a text and gauge. The memory utilization information is displayed in the text form of a used memory/total memory. Also, the memory utilization information is presented with a colored bar of the gauge.
  • the gauge bar can be placed in left or right side of the memory monitor item according the position of the memory monitor item on the screen and changed in color (for example, red) if the memory utilization percentage is over a predetermined percentage (for example, 95%). If a “go to memory status” command is input through a menu screen, a detailed memory utilization screen is displayed to provide information on the usage of the memory.
  • the clock item is a function item for alarming at preset times.
  • FIG. 9 g illustrates the clock item image in the normal mode. If a preset time has arrived, the clock item is animated as illustrated in FIG. 9 h. If the clock item is selected by the keypad manipulation, the clock item is presented on the screen in the form of an enlarged clock item image with the present time.
  • the preset time can be presented in the form of a text and image.
  • the preset time can be of a variety of functions such as “morning call,” “schedule,” “to do,”.
  • the text can be presented in various formats. For example, a text format 1 can be used for a title of the schedule and a text format 2 for date and time.
  • a default text “no title” is presented.
  • the time is presented in the form of TT/MM, and the date is presented in the form of today/tomorrow/a day of the week.
  • the text can be positioned at a left or right side of the clock item according to the position of the clock item on the screen. If multiple schedules are preset, the most closely arranged schedule is shown. If a “go to alarm status” command is input through a menu screen, a detailed schedule screen is displayed for informing preset alarm schedule details.
  • a structure and presentation of a function item are described with the mailbox item as an example hereinafter with reference to FIGS. 9 j to 9 l.
  • the mailbox item is a function item for informing the number of the unread messages among the received messages.
  • FIG. 9 j illustrates the mailbox item image in the normal mode. If a message is received in the normal mode, the mailbox item is animated as illustrated in FIG. 9 k. If the mailbox item is selected by the keypad manipulation, the mailbox item is presented on the screen in the form of an enlarged mailbox image with the information on the received and unread messages. The information on the received and unread messages can be presented with a text and image. The text can be presented
  • the text can be presented in various formats. For example, a text format 1 can be used for a category of the information (in this case, New Message), and a text format 2 for the numbers of the messages in the form of “new message/total message.”
  • the text can be positioned at a left or right side of the mailbox item according to the position of the mailbox item on the screen. If a “go to inbox” command is input through a menu screen, a message list screen is displayed for illustrating the received messages.
  • the motion items are items for making the screen presentation of the character house dynamic.
  • the motion items move around on the screen according to movement lines by item.
  • FIGS. 10 a to 10 c are views illustrating steps of presenting motion items in a screen image presentation technique according to an exemplary embodiment of the present invention.
  • the motion items of the character house include animals and movable objects as shown in FIG. 10 a. Each motion item moves in accordance with a unique movement line and moves at a preset animation time.
  • the motion items are animated on the basis of the position information contained in the character house information such that each motion item starts moving at the preset time along the preset movement line as illustrated in FIG. 10 b.
  • the motion item can be configured to move at a random speed and turn the progress direction in random probability.
  • the motion items when moving across each other, the motion items are overlapped.
  • the motion items are illustrated in an order of the presentation priorities (depth) that are preset for the respective items. All the motion items are assigned presentation priorities so that the motion item with the highest priority is presented when at least two motion items are overlapped while on the move. Some motion items have higher priorities than that of the character, and others have lower priorities.
  • the motion item “collie” (a dog) has a priority higher than the character such that the legs of the character are covered by the collie.
  • a special motion item in FIG. 10 a, a butterfly
  • FIG. 10 a a special motion item
  • the motion items are animated to move at a certain speed (for example, 5 pixels/sec) so they may disappear out of the character house (cross the border of the screen) and then re-appear. If a motion item encounters the character or other item while moving along the movement line, the motion item or the character with a higher priority is displayed on top. If a motion item with a priority lower than that of the character passes across the character, the motion item is shaded by the character.
  • a certain speed for example, 5 pixels/sec
  • FIGS. 11 a to 11 c are views illustrating steps in the presentation of decoration items in a screen image presentation technique according to an exemplary embodiment of the present invention.
  • FIG. 11 a illustrates exemplary decoration items that can be presented for decorating the character house
  • FIG. 11 b illustrates exemplary character houses decorated with the decoration items
  • FIG. 11 c illustrates specific decoration items shown in an enlarged form on the screen.
  • the decoration items to be applied to the character house include flowers, furniture, and other ornamental things. Such decoration items can be presented with the character in the character house as illustrated in FIG. 11 b. Some decoration items have special meanings.
  • the special decoration items such as a trophy and a clover shown in FIG. 11 c are obtained as a reward for winning a game or for a certain achievement of a specific state of the mobile phone, but bought in cash.
  • the decoration item “trophy” is used for informing a number of wins of the game, such that all wins of games may change the trophy in color and type and increases the number of the trophies.
  • the character quotient module 260 can be selectively set to be visible or invisible. All the elements constituting the character house are presented in association with the character house.
  • the character house also can be set to differently operate at a specific time or situation. For example, the character house can be configured such that the character gives greetings at specific times or when a specific event occurs.
  • the character gives the user greetings at the preset times and an item or character money can be given as a reward for the behavior.
  • FIG. 12 shows screen images illustrating character behaviors in the screen image presentation technique according to an exemplary embodiment of the present invention.
  • the greeting expression is displayed in the form of a text balloon and the greeting expression changes according to the time periods of the day.
  • the greeting behavior of the character can be configured such that the character greets at a predetermined number of times per day (such as, 4 times a day). If a determination is made that the character greets 4 times a day, 4 greeting time zones are configured: morning (08:30 ⁇ 09:30), afternoon (12:00 ⁇ 13:30), evening (17:30 ⁇ 18:30), and night (23:30 ⁇ 00:30). Accordingly, the character greets once at each greeting time zone. If the mobile phone is off or in character house disabled modes at each greeting time zone, the greeting behavior is disabled. Therefore, the mobile phone is not in the idle mode where the character house screen is presented (for example, when a clam shell type phone is unfolded or a slide type phone is opened), the greeting behavior function is disabled.
  • the greeting behavior When the greeting behavior starts, the greeting behavior remains during a predetermined time period (for example, 10 seconds). If a specific event (for example, an incoming call or a key input for activating a specific function) occurs while the character greets, the greeting is performed if the greeting behavior is maintained over a predetermined amount of time (for example, 3 seconds). If the greeting behavior stops before 3 seconds elapse and the character house is activated at the greeting times zone, the greeting behavior starts again.
  • the character is preferably associated with different greeting behaviors for the respective greeting time zones as in table 2. TABLE 2 Class Time Zone Text Morning 08:30 ⁇ 09:30 Hello! Afternoon 12:30 ⁇ 13:30 Good ⁇ Evening 17:30 ⁇ 18:30 Bye ⁇ Night 23:30 ⁇ 00:30 See U ⁇
  • the mobile phone can generate an event.
  • the event may occur in a probability set according to the times of the greeting behaviors. In this exemplary embodiment of the present invention, the probability is about 10%. If the event does not occur during the greeting time zone (that is, the mobile phone is off during the greeting time zone or the character house is not activated), the event will become invalid.
  • FIG. 13 is a screen image illustrating a character animation with an event in a screen image presentation technique according to an exemplary embodiment of the present invention.
  • a text is presented in the form of a text balloon.
  • the text expression as an even is presented in the text balloon.
  • the text balloon is displayed with a 30% transparency.
  • the text balloon for expressing the event can be presented together with an image.
  • a gift event text is presented with a gift image as shown in FIG. 13 .
  • the gift image is preferably presented at a bottom left or bottom light portion of the screen.
  • the gift event can be replaced by money or other item event.
  • the items and background are selected and arranged on the screen for decorating the character house.
  • FIG. 14 is a flowchart illustrating a character house presentation procedure of a screen image presentation method according to an exemplary embodiment of the present invention.
  • the controller 110 controls the operations of the mobile phone and determines whether the mobile phone enters a character house presentation mode (S 311 ). If the mobile phone enters the character house presentation mode, the controller verifies the character house configuration information (S 313 ).
  • the character house presentation mode can be identical with the idle mode of the mobile phone. In the idle mode, the controller 110 displays the character house on the screen of the display unit 143 during a predetermined display period. If the display period ends, the controller 110 turns off the display unit 143 to save power. If a clam shell or a slide of the mobile phone is open in the idle mode, the controller 110 turns on the display unit 143 .
  • the character house configuration information includes parameters configured using a setting application.
  • the parameters may include information for activating or deactivating the character house function.
  • the character house configuration information may include character house on/off information, character information display on/off information. If the character house configuration information includes the character house function related information, the controller 110 determines whether the character house function is turned on (S 315 ). If it is determined that the character house function is turned off, the controller 110 displays a normal idle mode screen.
  • the controller 110 verifies the character house information (S 319 ).
  • the character house information is on the currently set character house.
  • the character house information includes the information on the background and items associated with the character house.
  • the background information includes a background image ID
  • the item information includes item IDs and positions of the items on the screen (X, Y coordinate).
  • the character house can be displayed with one or more items, and the items are arranged in an order of the presentation priorities such that if two items are overlapped, the item with a higher priority is displayed on the other item.
  • the controller 110 recognizes the background and items with their positions on the screen of the display unit 143 .
  • the controller 110 controls to present the background, items, and character (S 321 , S 323 , and S 325 ).
  • the character house presentation is performed in the order of the background, items, and character.
  • the presentation order of the elements can be changed.
  • the elements of the character house can be presented in an order of the background, character, and items; an order of the character, background, and items; and an order of the character, items, and background. That is, the character house presentation can be performed in various orders.
  • the steps S 321 to S 325 the character house presentation is completed.
  • the mobile phone determines whether an event (for example, a function is selected by a key input) occurs or a predetermined display time ends (S 327 ). If an event occurs or the display time ends, the controller 110 ends the presentation of the character house and then performs the selected function or turns off the display unit. Also, it is possible to perform the selected function while maintaining the character house screen. For example, if a message is received while the character house is displayed or a message transmission mode is activated, the controller 110 can maintain the presentation of the character while reading the received message and writing a message to transmit. The controller 110 displays a message received or a message to be transmitted in a transparent message window on the character house screen. Also, the controller 110 can increase the transparency of the character house presentation and display the message on the character house presentation. Other events can be processed while the character house is presented in the similar manner.
  • an event for example, a function is selected by a key input
  • a predetermined display time ends the controller 110 ends the presentation of the character house and then performs the selected function or turns off the display unit
  • FIG. 15 is a flowchart illustrating the background presentation procedure of FIG. 14 .
  • the controller 110 verifies a background ID of the character house (S 361 ), and determines whether the background ID is 0 (S 363 ).
  • the background ID “0” means the character house does not use any background images. If the background ID is 0, the controller 110 loads a normal background image set as a default image (S 365 ). If the default background image is not set, no background image is presented.
  • the controller 110 retrieves a background corresponding to the background ID and presents the retrieved background on the screen. According to an exemplary implementation, the controller 110 selects one of the backgrounds as shown in FIG. 6 on the basis of the background ID.
  • the background of the character house can be represented by multiple images that are presented at different time periods.
  • the background of the character house can be represented by a fixed single image.
  • the controller 110 determines whether the retrieved background is a multiple image background (S 364 ). If the retrieved background is a multiple image background, the controller verifies the current time (S 367 ) and determines that the current time is greater than or equal to 06:00 and less that 12:00 (S 369 ). If the current time is greater than or equal to 06:00 and less than 12:00, the controller 110 selects a morning image assigned for the morning or afternoon from the multiple image background and then displays the morning image on the screen (S 371 ). If the current time is not greater than or equal to 06:00 or if the current time us not less than 12:00, the controller 110 determines whether the current time is greater than or equal to 12:00 and less than 18:00 (S 373 ).
  • the controller 110 selects an afternoon image assigned for the afternoon from the multiple image background and then displays the afternoon image on the screen (S 375 ). If the current time is not greater than or equal to 12:00 or not less than 18:00, the controller 110 determines whether the current time is greater than or equal to 18:00 and less than 23:59, or greater than or equal to 00:00 or less than 06:00 (S 377 ). If the current time is greater than or equal to 18:00 and less than 23:59, or greater than or equal to 00:00 or less than 06:00, the controller 110 selects an evening or night image assigned for the evening or night from the multiple image background and displays the evening or night image on the screen (S 379 ). At step S 364 , the retrieved background is not a multiple image background, the controller regards the retrieved image as a fixed single image background and displays the fixed signal image as the background (S 381 ).
  • a signal fixed image can be used by changing the brightness of the image for the respective time periods of a day.
  • FIG. 16 is a flowchart illustrating the item presentation procedure of FIG. 14 .
  • the item presentation procedure of FIG. 16 is described with the character information presentation of FIG. 7 b.
  • the character information presentation can be performed with the background presentation procedure, or the character presentation procedure of FIG. 19 .
  • the character information presentation is included in the item presentation procedure as shown in FIG. 16 .
  • the emoticons for expressing the emotion of the character, text for expressing the state of the character, and the character quotients, shown in FIG. 7 b, included in the character information are explained.
  • the controller 110 verifies whether character house configuration information includes information on a character quotient presentation function (S 401 ) and determines whether the character information presentation function is on (S 403 ). If the character information presentation function is on, the controller 110 verifies the character quotient information (S 405 ) and presents a character emoticon on the basis of the character quotient information in the emotion icon display region 263 of FIG. 7 b (S 407 ).
  • the emotion of the character can be represented by 5 emotion modes: happy, normal, angry, sad, and tired.
  • the emoticon to be displayed in the emotion icon display region 263 can be one of the emoticons corresponding to the emotion modes, as shown in FIG. 8 .
  • the controller 110 presents a text corresponding to the current state of the character in the status display region 265 of FIG. 7 b (S 409 ) and then presents the character quotients (SQ, HQ, IQ, and EQ) in the quotient information display region 267 of FIG. 7 b in the form of bar graphs (S 411 ).
  • the values of character quotients are presented as different colored bars.
  • the controller 110 verifies the character quotients and determines the emotion mode on the basis of the character quotients. If the emotion mode is determined, the controller 110 presents an emoticon corresponding to the emotion mode, a text appropriate for the state of the character, and the quotients in the form of the bar graphs.
  • the character information based on the character quotients can be presented at a fixed position in the character house (for example, a top left portion) as shown in FIG. 7 a in the form of FIG. 7 b.
  • the character quotient information cannot be presented according to the user's configuration.
  • the user decides whether to present the character information. If the user does not activate the character information presentation function, the controller recognizes the configuration at step S 403 such that the steps S 405 to S 411 are skipped. According to an exemplary implementation, the character information based on the character quotients information shown in FIG. 7 b is not presented.
  • the controller verifies the information on an item with a lowest priority (depth) (S 413 ).
  • the item information includes an item ID and display position.
  • the item can be one of the motion items, function items, and decoration items.
  • two or more items can be overlapped.
  • the items are presented in an order of the presentation priorities assigned to the items such that the item with a lowest priority is presented first. That is, when two or more items are overlapped, only the item with the highest priority is completely presented. If the item on top is disabled, the item with the second highest priority is presented.
  • the controller 110 determines whether the item information exists (S 415 ). If the item information exists, the controller 110 determines whether the item is a motion item on the basis of the item information (S 417 ). If the item is a motion item, the controller 110 presents the motion item (S 419 ) and verifies another item with the next lowest priority (S 421 ). If it is determined that the item is not a motion item at step S 417 , the controller 110 determines whether the item is a function item (S 423 ). If a determination is made that the item is a function item, the controller 110 presents the function item (S 425 ). If the item is not a function item, the controller 110 presents the item as a decoration item (S 427 ).
  • the item presentation is performed in a different manner based on the type of item. That is, the motion item, function item, and decoration item are presented in different procedures.
  • FIG. 17 is a flowchart illustrating a procedure for presenting a motion item in the screen image presentation technique according to an exemplary embodiment of the present invention.
  • the controller 110 verifies the motion item information on the basis of the item ID (SS 431 ).
  • the motion item information includes an item ID, a sojourn time for staying the item at the initial position, a coordinate represented by a horizontal distance (X axis distance) for moving the item to the next position and a vertical distance (Y axis distance) for moving the item to the next position, and initial position information.
  • the initial position information contains one of 8 coordinates listed in table 3.
  • the motion item is assumed to start moving from an edge of the screen as the initial position (such as 4 sides and 4 corners of the screen of the display unit 143 ) in table 3.
  • the motion item may also be configured to start moving at a random position in the screen of the display unit 143 . TABLE 3 1.
  • the controller 110 After verifying the motion item information, the controller 110 obtains the initial position of the motion item to animate the motion item at the initial position (S 433 ). The animation is maintained before the sojourn time is ended and then the motion item is positioned at the next position through the steps S 435 to S 445 . After animating the motion item, the controller 110 determines whether the sojourn time expires (S 435 ). If the sojourn time is expired, the controller 110 calculates a coordinate of the next position to which the motion item moves (S 437 ). After calculating the next position coordinate, the controller 110 determines whether the calculated next position is in a available display range on the screen of the display unit 143 (S 439 ).
  • the controller 110 moves the motion item to the next position (S 441 ) and then returns to step S 433 to animate the motion item at the next position.
  • the controller 110 determines that the movement of the motion item is completed and moves positions the motion item at the initial position (S 443 ) and repeats the step S 433 if the calculated next position is not in the available display range.
  • the motion item is configured with one of the initial positions described in table 3 and moves to a predetermined coordinate on the screen after a predetermined period of time.
  • the motion item is animated at the initial position, moved to the next position after the sojourn time, and then animated again, under the control of the controller 110 .
  • the movement time is determined on the basis of the sojourn time and a distance between the two positions. If the sojourn time is short and the distance between the two positions is close, the motion item moves fast. Alternatively, if the sojourn time is long and the distance between the two positions is great, the motion item moves slowly.
  • the movement direction of the motion item depends on the relationship of the two positions (coordinate in the screen).
  • the motion item can move in horizontal, vertical and diagonal directions and can be determined in consideration of the distance between the two positions.
  • the motion item can be one of the items depicted in FIG. 10 and move in the manner illustrated in FIG. 10 b.
  • the motion items move on the screen as illustrated in FIG. 10 b in accordance with the procedure of FIG. 17 .
  • the movement of the motion item stops once the character house presentation ends.
  • the movement of the motion item may also be set to stop at a time different from the character house presentation end time. That is, the motion item presentation time can be shorter than the character house presentation time. According to an exemplary implementation, the motion item disappears before the character house presentation stops.
  • the controller 110 maintains the animation of the motion item (S 445 ) and determines whether the motion item presentation time is ended, (S 447 ). If the motion presentation time is ended, the controller stops presentation of the motion, and otherwise, repeats the step S 435 .
  • FIGS. 18 a to 18 d are flowcharts illustrating a procedure for presenting function items in the screen image presentation technique according to an exemplary embodiment of the present invention.
  • the controller 110 verifies a type of function item (S 452 ).
  • a type of function item S 452 .
  • three function items are assumed to exist. This items are the mailbox item, alarm clock item, and memory monitor item. The number and type of the function items can be changed.
  • the controller 110 determines whether the function item is a mailbox item (S 453 ). If the function item is the mailbox item, the controller verifies a presentation position of the mailbox item and presents mailbox item at the presentation position on the screen (S 455 ). Next, the controller 110 verifies a number of received but unread messages (S 457 ) and determines whether the number of unread messages are greater than 0 (S 459 ). If the number of the unread messages is not greater than 0, the controller presents an empty mailbox image.
  • the controller 110 displays the number of the unread messages with the mailbox image (S 461 to S 467 ).
  • the number of the unread messages can be presented in various manners.
  • the number of unread messages can be displayed on the mailbox image.
  • the number of unread messages can be displayed by means of a text balloon.
  • the text balloon has a shape of a postcard with the number of the unread messages.
  • the mailbox item presentation is explained with the text balloon image for displaying the number of the unread messages.
  • the controller 110 determines an unread message number presentation position.
  • the controller 110 verifies the left and right side spaces of the mailbox items and arranges the text balloon at the left or right side. If a determination is made that the text balloon is arranged at the right side, the controller 110 determines whether the text balloon presentation position is in the available display range of the screen (S 463 ). If the text balloon presentation position is in the available display range, the controller 110 presents the text balloon at the right side of the mail box (S 467 ). If the text balloon presentation position is not in the available display range, the controller 110 determines to the text balloon at the left side of the mailbox item (S 465 ) and repeats the step S 463 .
  • the controller determines whether the function item is an alarm clock item (S 469 ) in FIG. 18 b. If the function item is the alarm clock item, the controller 110 verifies scheduled alarms on the basis of alarm information (S 471 ). The alarm information includes the presentation position of the alarm clock item on the screen of the display unit 143 . After verifying the scheduled alarms, the controller determines whether the there is scheduled alarms (S 473 ). If no alarm is scheduled, the controller 110 presents an empty alarm image. If there is at least one scheduled alarm, the controller 110 retrieves a next scheduled alarm and displays an alarm clock image with the next scheduled alarm (S 475 ).
  • the alarm clock item can be implemented with a clock image. According to an exemplary implementation, the alarm clock image may have the needles arranged to show the scheduled alarm time.
  • the controller 110 verifies an alarm time presentation position and determines whether there is enough space to present the alarm time at the right side of the alarm clock item in the screen. If there is enough space to present the alarm time, the controller 110 presents the next scheduled alarm time at the right side of the alarm clock item (S 477 ). If it is determined that there is not enough space to present the alarm time at the right side of the alarm clock item (S 479 ), the controller 110 presents the alarm time at the left side of the alarm clock item ( 483 ).
  • the alarm can be represented by a morning call, a meeting schedule, and a reservation, among others.
  • the scheduled alarms are activated at the scheduled times with the specific events.
  • the alarm time can be shown in the form of YYYY/MM and HH/MM as a text.
  • the alarm time text can be placed at the left or right side of the alarm clock item according to the position of the alarm clock item.
  • the alarm time of the next closest alarm is presented.
  • the alarm time can be presented by means of the alarm clock image.
  • When the alarm time is presented by means of the alarm clock image.
  • the alarm time text may not be presented.
  • the controller 110 determines that the function is a memory item. Next, the controller 110 verifies a presentation position of the memory item from the character house information and presents the memory item at a verified presentation position (S 484 ). After presenting the memory item on the screen, the controller 110 verifies a total memory capacity and used memory capacity (S 485 ) and then displays the information on the total memory capacity and the user memory capacity (S 487 to S 493 ). That is, after verifying the memory information, the controller 110 verifies the side of the memory item that has enough space for presenting the memory information.
  • the controller 110 presents the memory information at the right side of the memory item.
  • the controller 110 presents the memory information at the left side of the memory item.
  • the memory item may inform of the memory information of the memory unit 130 of FIG. 1 and the memory information can be shown in the form of a gauge bar.
  • the gauge bar can be located at the left or right side according to the position of the memory item on the screen. The gauge bar shows the memory utilization and changes in red if the memory utilization reaches over 90%.
  • a function item is presented in such a manner that the controller verifies the presentation position of the function and presents the function item at the presentation position on the screen of the display unit 143 together with the function information obtained by the function item.
  • the function information is presented in such a manner that the controller verifies which side of the function item has enough space for presenting the function information and then presents the function information at a left or right side of the function item, where there is enough space.
  • the function information can be shown as a text or a graphical gauge. When the function information is provided in the text, the graphical gauge may or may not be used.
  • FIGS. 18 c and 18 d are flowcharts illustrating a procedure for presenting function items in the screen image presentation technique according to another exemplary embodiment of the present invention.
  • the controller 110 verifies a type of function item (S 751 ).
  • the function items include a mailbox item, alarm clock item, and memory monitor item. The number and types of the function items can be changed.
  • the controller presents the function item in the form of an animated function item for attracting the user's interest and presents the function information in a popup window during a popup display time.
  • the function items are presented in the following order: mailbox item, alarm clock item, and memory item.
  • the controller 110 determines whether the function item is the mailbox item (S 753 ). If the function item is the mailbox item, the controller 110 verifies an amount of the unread mail (S 755 ) and then determines whether the amount of the unread mail is greater than 0 (S 757 ). If the amount of the unread mail is 0, the controller 110 presents an empty mailbox image on the screen of the display unit 143 as shown in FIG. 9 j (S 767 ) to notify that there is no unread message.
  • the controller 110 presents an animated mailbox image as shown in FIG. 9 k. After presenting the animated mailbox image, the controller determines whether a popup time is arrived (S 761 ). If the popup time has arrived, the controller 110 presents a popup window as shown in FIG. 9 l (S 763 ). The popup window is displayed during a predetermined popup time and then disappears. During the presentation of the popup window, the controller 110 simultaneously starts a popup counter. After the popup window is presented, the controller determines whether the popup counter is expired (S 765 ). If the popup counter is expired, the popup window disappears.
  • the controller 110 determines whether the function item is an alarm clock item (S 771 ). If the function item is the alarm clock item, the controller 110 verifies scheduled alarms (S 773 ) and determines whether there are scheduled alarms (S 775 ). If there is no scheduled alarm, the controller 110 presents an empty alarm image on the screen of the display unit 143 for notifying that no scheduled alarm exists as shown in FIG. 9 g (S 785 ). If there is at least one scheduled alarm, the controller 110 presents an animated alarm clock image as shown in FIG. 9 h (S 777 ) and determines a popup time has arrived (S 779 ). If the popup time is arrived, the controller 110 presents a popup window as shown in FIG.
  • the popup window is displayed during a predetermined popup time and then disappears.
  • the controller 110 simultaneously starts a popup counter. After the popup window is presented, the controller 110 determines whether the popup counter is expired (S 783 ). If the popup counter is expired, the popup window disappears.
  • the popup window presentation time may be configured in consideration of the alarm time. According to an exemplary embodiment of the present invention, the scheduled alarm is assumed to be an alarm set with a specific date and time for “meeting” and “to do” rather than an everyday occurrence alarm (for example, morning call), and displays alarms before 6 hours.
  • the controller 110 determines whether the function item is a memory monitor item (S 786 ). If the function item is the memory monitor item, the controller 110 verifies memory utilization (S 787 ) and determines whether the memory utilization is in a predetermined normal range (S 789 ). If the memory utilization is in the normal range, the controller 110 presents only the memory monitor item image on the screen of the display unit 143 for notifying that the memory utilization is normal, as shown in FIG. 9 d (S 799 ). If the memory utilization exceeds the normal range, the controller 110 presents an animated memory monitor as shown in FIG. 9 e (S 791 ) and determines a popup time has arrived (S 793 ).
  • the controller 110 If the popup time is arrived, the controller 110 presents a popup window as shown in FIG. 9 f (S 795 ).
  • the popup window is displayed during a predetermined popup time and then disappears.
  • the controller 110 simultaneously starts a popup counter.
  • the controller 110 determines whether the popup counter is expired (S 797 ). If the popup counter is expired, the popup window disappears.
  • the memory monitor item can be displayed regardless of the memory utilization. According to an exemplary implementation, the controller 110 verifies the memory utilization and then presents the memory utilization in the form of a popup window.
  • the function items can be presented together with the character house.
  • the function items are displayed in a preset order. While presenting the function items in the preset order, the controller determines whether to present the function item in an active state. If it is required to present the function item in the activate state, the controller 110 presents the function item as an activated function item image and then provides the information related to the function in the form of a popup window after a while. The popup window is presented during a predetermined popup time.
  • the controller 110 presents in the order of the mailbox item as an animated mailbox image followed by the mailbox popup window, the alarm clock item as an animated alarm clock image followed by the alarm popup window, and the memory monitor item as an animated memory monitor image followed by the memory utilization popup window.
  • the function items are presents in the character house as shown in FIGS. 9 j, 9 g, and 9 d after the popup windows disappear.
  • the decoration items are presented on the basis of the decoration item information contained in the character house information.
  • the item information includes an item ID that provides information on the type of the decoration item and a presentation position on the screen.
  • the controller 110 retrieves a decoration item image corresponding to the item ID and presents the retrieved decoration item image at the presentation position on the screen as shown in FIG. 11 a.
  • the decoration item is used for improving aesthetic factor of the character house, regardless of any function.
  • the decoration item can be added or removed according to the user's configuration.
  • the character In the character house, the character is implemented with three behavior properties.
  • the character greets the user of the mobile phone at preset times.
  • the character irregularly gives the user a gift while greeting the user.
  • the character expresses a state of the mobile phone or an emotion reflecting the state of the mobile phone in an animation or a motion.
  • the mobile phone has various emotions that represent a user's contribution to the mobile phone.
  • the character house can be presented only with the third behavior property. That is, the greeting and gift-giving behavior properties can be disabled when the character house is presented.
  • the greeting, gift-giving, and animation behavior properties are enabled.
  • the character house can consist of the items and background as well as the character, and the character house is presented in a specific operation mode of the mobile phone.
  • the character house can be shown as an idle mode screen (for example, when a clam shell of the mobile terminal is unfolded or a slide mobile phone is slid open). If an event occurs while the character house is presented, the character house disappears. In other words, the idle screen is replaced by other function screen.
  • the character house can be maintained when a specific event occurs (for example, a message processing event).
  • FIG. 19 is a flowchart illustrating a character presentation procedure of the screen image presentation method according to an exemplary embodiment of the present invention.
  • the controller 110 verifies a current time (S 501 ) and then determines whether the current time is in any of preset greeting behavior time zones (S 503 , S 513 , S 517 , and S 521 ).
  • the greeting behavior time zones are set to four time zones as in the table 2. The greeting behavior is shown once in a greeting behavior time zone.
  • the controller 110 determines whether the current time zone is in the greeting behavior time zones (S 503 , S 513 , S 517 , and S 521 ). If the current time zone is in the greeting behavior time zones, the controller 110 determines the greeting behavior is shown (S 505 , S 515 , S 519 , and S 523 ).
  • the controller 110 skips the greeting behavior and controls the character to be in a normal behavior (S 511 ).
  • the controller 110 records the current time (S 507 ) and controls the character to show the greeting behavior (S 509 ). Recording the current time zone helps to avoid situations in which the greeting behavior is not performed in the same greeting behavior time zone. After the greeting behavior is performed, the controller 110 controls the character to be in the normal behavior (S 511 ).
  • the greeting behavior presentation procedure is performed as shown in FIG. 20 .
  • the greeting behavior is presented as the character greets to the user and can be performed with a salutation and an addition event for giving an item or character money in a probability.
  • FIG. 20 is a flowchart illustrating a greeting behavior presentation procedure of FIG. 19 .
  • the controller 110 verifies a current greeting time zone (S 531 ) and verifies current character information (S 533 ).
  • the controller 110 controls the character to show the greeting behavior in the form of an animation (S 535 ) and retrieves a greeting text corresponding to the current greeting time zone (S 537 ).
  • the controller 110 presents the retrieved greeting text in the form of a text balloon (S 539 ).
  • the greeting behavior can be presented as shown in FIG. 12 .
  • the greeting text can be presented in the form of the text balloon of FIG. 12 .
  • the greeting text can be changed, depending on the greeting time zones.
  • the controller 110 determines that the greeting behavior has been performed if the greeting behavior is performed for a tolerable time.
  • the tolerable time is assumed to be 3 seconds.
  • the controller 110 determines whether the greeting behavior is continued over the tolerable time (in this embodiment, 3 seconds) (S 541 ). If the greeting behavior is not continued over the tolerable time, the controller 110 maintains the greeting behavior and text balloon presentation (S 557 ). While maintaining the greeting behavior and the text balloon presentation, the controller determines an event occurs such that the character house presentation is terminated in 3 second-tolerable time (S 559 ). If an event occurs in tolerable time, the controller 110 stops the greeting behavior and then returns. According to an exemplary implementation, the greeting behavior is performed again, if the character house is activate in the same greeting behavior time zone.
  • the controller determines that the greeting behavior is completely performed and records the completion of the greeting behavior in the greeting behavior time zone (S 543 ).
  • the controller 110 determines whether the event presentation time has arrived (S 545 ). The event may occur in a random probability set according to the times of the greeting behaviors.
  • the controller 110 If the event presentation time has arrived, the controller 110 presents the greeting event (S 547 ).
  • FIG. 21 is a flowchart illustrating a greeting event presentation procedure that occurred during a greeting behavior of the character of the screen image presentation technique according to an exemplary embodiment of the present invention.
  • the greeting event can randomly occur while the character performs the greeting behavior.
  • the greeting event may also be presented in the same manner with the character's greeting behavior except for the screen layout and the text.
  • the controller 110 selects an item or character money as a gift at the greeting event presentation time (S 561 ) and presents an image corresponding to the item or the character money to give as the gift (S 563 ).
  • the controller 110 verifies the information of the character (S 565 ) and animates the character (S 567 ).
  • the controller 110 presents the greeting text in the form of the text balloon (S 569 ).
  • the controller 110 When presenting the character according to the greeting event, the controller 110 animates the character with the text and/or sound as shown in FIG. 13 .
  • the emotion mode of the character is fixed to “happy” and the character is presented with the item image and/or the character money image as the gift.
  • the item image and the character money image provided in the greeting event presentation procedure can be displayed at the bottom left or bottom right portion of the character house as shown in FIG. 13 , and the text expression related to the gift is presented in the form of a text balloon having a 30% transparency.
  • the controller 110 determines whether 10 seconds are elapsed after the presentation of the greeting expression text balloon (S 549 ). If 10 seconds have elapsed, the controller 110 makes the text balloon and the character animation image disappear (S 551 ).
  • the controller maintains the presentation of the current character animation and the greeting expression text balloon (S 553 ).
  • the greeting behavior of the character is assumed to be maintained for 10 seconds. Accordingly, before 10 seconds have elapsed during the presentation of the greeting behavior of the character, the controller 110 maintains the greeting behavior of the character (S 549 , S 553 , and S 555 ). If 10 seconds have elapsed for the start of the greeting behavior of the character, the controller 10 makes the greeting expression text balloon and the character animation image disappear (S 551 ).
  • the controller 110 controls the character of the character house to show the greeting behavior corresponding to the greeting time zone.
  • the greeting behavior is for greeting to the user of the mobile phone and is performed on in every greeting time zone. If the mobile phone is turned off or the character house is not activated, the greeting behavior of the character is skipped.
  • the greeting behavior of the character is performed in a random probability to give the user an unexpected gift.
  • the gift-giving behavior is performed during the greeting behavior time.
  • the gift-giving behavior may be performed at any time during the presentation of the character house. If the gift-giving behavior is performed at any time during the character house presentation, the controller 110 can control the gift-giving behavior to be performed when the character is presented.
  • the controller 110 controls the behavior of the character so that it is normal (S 511 ) after performing the greeting behavior of the character out of the greeting time zones or after performing the greeting behavior.
  • the character behaviors are presented in the procedure of FIG. 22 a or FIG. 22 b.
  • FIGS. 22 a and 22 b are flowcharts illustrating an animation behavior presentation procedure of the screen image presentation technique according to an exemplary embodiment of the present invention.
  • the character is moving and animated together with a text and sound for expressing the state and emotion of the character while the character house is presented.
  • the character is animated after the text and sound expression is presented in the character house.
  • the character behavior can be vividly expressed in the character house in various manners.
  • the character can be represented by a plurality of animation images such that the animation images are continuously displayed to form an animation of the character.
  • the character can be presented to move along a preset movement line.
  • the character animation is implemented by presenting the animation images in series at regular intervals, similar to presenting the function items as shown in FIG. 9 e, 9 h, and 9 k.
  • the character movement is implemented by moving the character on the basis of a preset movement direction, movement distance, and movement time. In order for the character to seamlessly move, the movement can be implemented with a plurality of images.
  • FIG. 23 shows a procedure for presenting a character movement.
  • FIG. 22 a shows a character movement behavior presentation procedure
  • FIG. 22 b shows a character gesture behavior presentation procedure
  • the controller 110 verifies information on the current character (S 571 ) and determines an initial appearance position of the character (an X, Y coordinate in the screen of the display unit 143 ).
  • the initial appearance position of the character can be randomly determined in a display range of the screen of the display unit 143 .
  • the controller 110 presents the character with 100% transparency (S 575 ).
  • the 100% transparency produces character invisibility so that the background and items images are displayed on the screen of the display unit 143 as shown in FIG. 4 a.
  • the controller 110 decreases the transparency of the character image at a predetermined rate so that the character gradually appears in the character house.
  • the controller 110 controls such that the character appears gradually by repeating the steps S 577 and S 579 . If the transparency of the character image is assumed to decrease by 10%, the controller 110 controls to gradually reduce the transparency of the character image to be opaque (alpha blending effect). Accordingly, the transparency of the character image decreases from 100% to 0% by 10% such that the character appears in the character as shown in FIG. 4 b.
  • the controller 110 outputs a sound effect representing the state of the character as shown in FIG. 4 c (S 581 ), animates the character as shown in FIG. 4 d (S 583 ), presents a text expressing the state and emotion of the character in the text balloon during a predetermined presentation time as shown if FIG. 4 e, and then terminates the animation of the character.
  • the presentation time is set to 3 seconds.
  • the controller 110 determines a next position of the character and then performs movement of the character (S 589 , S 597 , and S 559 ). In order to present the character movement, the controller determines position shift information (S 587 ). The position shift information is determined according to a procedure as shown if FIGS. 23 a and 23 b.
  • FIGS. 23 a and 23 b show the character shift information determination procedure in detail.
  • the controller 110 verifies an initial appearance position of the character (S 601 ) and determines a movement type of the character (S 603 ).
  • the movement type includes a vertical movement, a horizontal movement, and a diagonal movement.
  • the movement type is randomly selected among the three movement types at step S 603 . It is assumed that the character moves by 5 pixels.
  • step S 615 If it is determined, at step S 615 , that the movement type is not the horizontal movement, the controller 110 regards the movement type as the diagonal movement.
  • the diagonal movement includes a top-right to bottom-left movement, a top-left to bottom-right movement, a bottom-right to top-left movement, and a bottom-left to top-right movement.
  • the controller 110 verifies the appearance position and movement type of the character and calculates the position shift information on the basis of the appearance position and the movement type of the character.
  • the movement direction of the character is changed according to the character's initial appearance position and the movement type of the character is randomly selected among the three movement types (the vertical, horizontal, and diagonal movement). The movement direction is determined such that the character secures the longest length to move at an axis of the center of the screen.
  • the controller 10 controls such that the character moves in the movement direction and according to movement type (S 589 to S 599 ).
  • the appearance position shift is performed in a predetermined interval. It is assumed that the character moves 20 pixels per second.
  • the character's appearance position shift is performed by 5 pixels every 0.25 seconds.
  • the character is moving and animated at the same time. That is, the character moves in a direction determined at every predetermined times by a unit of predetermined number of pixels.
  • the character can be animated while moving or not moving.
  • the controller 110 determines whether the position shift time arrived (S 589 ). If the position shift time has not arrived, the controller 110 animates the character (S 597 ).
  • the controller 110 calculates a next position of the character (S 591 ).
  • the next position is obtained according to the procedure shown in FIGS. 23 a and 23 b.
  • the controller 110 determines whether the next position of the character is in a display range of the screen of the display unit 143 (S 593 ). If the next position of the character is in the display range of the screen, the controller 110 controls to move the character to the next position and animate the character.
  • the controller 110 moves the character on the screen of the display unit 143 in the predetermined movement direction at the predetermined movement times by unit of predetermined number of pixels as shown in FIGS. 4 f and 4 g.
  • the controller 110 recognizes the deviation of the character and returns to the step S 571 to show the idle mode screen as shown in FIG. 4 h. Consequently, the presentations of FIGS. 4 b to 4 g are repeated.
  • the above operation is repeatedly performed while presenting the character house. If the presentation of the character ends (S 599 ), the controller 110 terminates the behavior of the character.
  • the movement of the character is performed after the text expression of the character disappears as shown in FIGS. 4 f and 4 g.
  • the movement direction of the character is randomly selected among the vertical, horizontal, and diagonal direction. That is the movement direction is determined such that the character secures the longest length to move at an axis of the center of the screen and the movement speed is determined at 20 pixels per second. If the character moves across the border of the screen of the display unit 143 to disappear, the character house is presented with the character. After about 2 seconds have passed, the operation depicted in FIG. 4 b is performed.
  • the presentation of the character behavior in FIGS. 22 a and 22 b is performed by repeating the operations of FIGS. 4 a to 4 g.
  • the character can be presented in 6 steps.
  • the first is an invisible character step as shown in FIGS. 4 a and 4 h, in which the character is not presented yet such that the character house is displayed with the items and background.
  • the second is a character presentation step, in which the character appears at a random position in the character house.
  • the initial appearance position of the character is close to the border of the screen of the display unit 143 such that the whole image of the character cannot be shown, the character is not presented.
  • the character can appear in a fade-in effect by decreasing the transparency of the character image. The transparency decreased from 100% to 0% (alpha blending image).
  • the third is a sound effect output step as shown in FIG. 4 c, in which a unique sound effect is output according to the state of the character after the character is completely presented. The sound effect is output once.
  • the fourth is an animation step as shown in FIG. 4 d, in which the character is animated after the sound effect is output. The character animation is repeated until the character moves to disappear out of the screen of the display unit 143 .
  • the fifth is a text presentation step as shown in FIG. 4 e, in which the state and emotion of the character are presented in text.
  • the text can be presented in the form of a text balloon as shown in FIG. 4 e.
  • the text balloon is presented with a semitransparent balloon with a 30% transparency in consideration of the time taken to read the text.
  • the text balloon disappears after about 3 seconds.
  • the sixth step is a character movement step as shown in FIGS. 4 f to 4 g, in which a movement direction of the character is randomly determined. After the movement direction and distance is determined, the character moves according to the movement direction and distance. The character moves to disappear out of the screen of the display unit 143 .
  • the character Since the character appears at a random position, the character may appear at different position at every appearance time.
  • the movement direction is randomly determined, the character may move in different directions at every movement time.
  • the sound effects and the texts can be differently prepared for expressing various states and emotions of the character. Since the appearance position, movement direction, sound effect, and text are presented differently at every presentation timings, various character behaviors can be provided to the user.
  • FIGS. 4 a to 4 g the operations illustrated in FIGS. 4 a to 4 g are performed and then the operations of FIGS. 4 c and 4 e can be skipped after the character moves to disappear out of the screen of the display unit 143 .
  • the sound effect and the text balloon appear once and the character appears at a random position and moves in a random direction.
  • the controller 110 verifies information of the current character (S 651 ) and determines a character appearance position (S 653 ).
  • the character appearance position can be a preset position on the screen of the display unit 143 .
  • the character appearance position is preset at a bottom right portion of the screen relative to the center of the screen of the display unit 143 .
  • the controller 110 presents the character with a 100% transparency (S 655 ). Since the transparency is 100%, the character is invisible. Accordingly, the character house is displayed only with the background image and item images as shown in FIG. 4 i.
  • the controller 110 decreases the transparency of the character at a predetermined rate such that the character appears gradually. That is, the controller 110 repeats the steps S 657 and S 659 for gradually presenting the character. If it is assumed that the transparency of the character image decreases 25% every 0.5 seconds, the character is completely presented in 2 seconds (alpha blending effect). The transparency of the character image decreases in an order of 100%, 75%, 50%, 25%, and 0% such that the character house is displayed as shown in FIG. 4 j. After the character has completely appeared, the controller 110 outputs a sound effect corresponding to the state of the character as shown in FIG.
  • the text presentation time can be set to about 1.5 seconds in consideration of the time taken for reading the text.
  • the controller 110 presents a character animation at steps S 665 to S 669 .
  • the character animation is repeated until the character disappears from the character house.
  • the character presentation is performed in a specific order according to the character appearance, output of sound effects, presenting text balloon, and character animation.
  • the screen image presentation method of an exemplary embodiment of the present invention presents the character house (screen) in a specific operation (idle mode) of the mobile phone.
  • the controller of the mobile phone presents a character house by arranging a background and items in the character house and presents a character.
  • the character expresses a state and emotion in the forms of a sound effect and text.
  • the character moves in random direction and shows an animation on the screen of the display unit 143 .
  • the background image is used as a background of the house and can be replaced with another image to graphically illustrate the progress of the time using character house.
  • the items presented with the character house includes movement items, decoration items, and function items.
  • the function items are used for improving user convenience by showing system information of the mobile phone.
  • the decoration items are used for improving the aesthetic effect of the screen of the mobile phone.
  • the motion items are used to make dynamic the screen by moving around the screen house.
  • the character house can be implemented with a special expression function such that the character gives special information at predetermined time zones or greets to the user.
  • the character house can be implemented such that the character gives an item or character money as a gift at unexpected timing while the character greets, resulting in user's enjoyment.
  • the character house is structured by presenting the elements in an order of the background, items, and character. However, the presentation order can be changed.
  • the character house is configured such that the greeting and event presentation are performed when the item is presented, however, the greeting and event presentation can be performed when the background or character is presented.
  • FIG. 24 is a flowchart illustrating a screen house presentation procedure of a screen image presentation method for a mobile phone according to another exemplary embodiment of the present invention.
  • the mobile phone In association with the character house presentation, the mobile phone can be in a character house configuration mode, character house editing mode, or character house presentation mode.
  • the character house configuration mode and character house editing mode can be accessed by a corresponding menu selection or a specific key (including a soft key) manipulation, and the character house presentation mode activates for a specific operation mode of the mobile terminal when the character house presentation function is enabled.
  • the controller 110 detects the execution of the character house configuration mode (S 719 ) and then displays a character house configuration menu screen for allowing a user to configure the character house presentation options (S 721 ).
  • the character options include a character house presentation on/off option and character quotients presentation on/off option. If the character house presentation on/off option is set to off, the controller 110 disables the character house presentation function. If the character quotients presentation on/off option is set to off, the controller 110 disables the presentation of the character quotients shown in FIGS. 7 b and 8 .
  • the character house option may include a greeting function on/off option. If the greeting function on/off option is set to off, the controller 110 disables the greeting behavior presentation shown in FIG. 12 and greeting event presentation shown in FIG. 13 .
  • the controller 110 detects the executing of the character house editing mode (S 715 ) and then displays a character house editing menu screen for allowing the user to edit the character house (S 717 ).
  • the character house editing mode it is possible to change, add, and delete the background, items, and character.
  • FIG. 25 is a state diagram illustrating a character house design procedure of a screen image presentation method according to an exemplary embodiment of the present invention.
  • the controller 110 displays a character house configuration screen in the character house configuration mode 810 , in which it is allowed to delete a character house or to set a default character house. If a “new character house” option is selected or a “character house editing” option is selected on the character house configuration screen, the controller 110 displays a character house editing screen in the character house editing mode 820 . From the character house editing mode 820 , it is possible to enter modes 830 to 870 for editing a background and items.
  • FIG. 26 is a state diagram illustrating transitions between the character house configuration mode and the character house editing mode of FIG. 25 .
  • FIG. 27 is a screen image illustrating a character house configuration screen in the character house configuration mode of FIG. 25 .
  • the character house configuration screen displays character houses that can be used for a specific mode screen image.
  • the mobile phone includes a plurality of character houses. These character houses are displayed on the character house configuration screen in the form of tiled windows as shown in FIG. 27 .
  • the character house configuration screen is assumed to display four character houses at once. If there are three character houses stored in the mobile phone, one of the four tiled window is shown empty. In relation with the empty character house window, it is impossible to set a default character house, delete a character house, and edit a character house. Also, a new character house can be put into the empty window. If an OK key is input while the empty window is selected, a character creation function is enabled. If the OK key is input while one of the tiled windows having the character houses is selected, the character house is set for a screen image for a specific operation mode of the mobile phone.
  • the cursor navigates on the tiled windows in accordance with a direction key (navigation key) manipulation.
  • the controller 110 displays a menu having options shown in table 4. TABLE 4 Empty window 2. new character house Non-empty window 1. set default 2. edit character house 3. delete character house
  • FIG. 28 is a state diagram illustrating transitions between the character house configuration mode and the character house editing mode of FIG. 25
  • FIG. 29 shows screen images illustrating a procedure for setting a “character house presentation on/off” option on the character house configuration screen of FIG. 25 .
  • the “character house presentation on/off” option can be set to “on” or “off.”
  • the controller 110 displays the character houses on the character house configuration screen in the form of tiled windows as shown in a first screen image of FIG. 29 . If one of the windows is selected while the “character house presentation on/off” option is set to off, the controller 110 highlights the selected window and displays an announcement message asking whether to use the character house presentation function as shown in a second screen image of FIG. 29 . If a user command is input for using the character house presentation function, the controller 110 enables the character house presentation function and displays the character house of the selected window as shown in a third screen image of FIG. 29 .
  • the controller 110 displays an announcement message asking whether to use the character house presentation function and then enables the character house presentation function according to the user command for using the character house presentation function.
  • FIG. 30 is a state diagram illustrating transitions among the character house editing mode and specific item editing modes of FIG. 25
  • FIG. 31 shows a screen image illustrating an example character house editing screen in the character house editing mode of FIG. 25 .
  • the character house editing screen is displayed when a “new character house” option or “edit character house” option is selected on the character house configuration screen.
  • the character house editing screen provides a menu with options for editing the background and item (1. Background, 2. New item).
  • a selected background is basically presented without items.
  • the character house editing screen displays a key input guide as shown in FIG. 31 . If a menu option is selected, the menu with options for editing the background and items is displayed.
  • FIG. 32 is a state diagram illustrating transitions between the character house editing mode and the background selection mode of FIG. 25 .
  • FIG. 33 shows screen images illustrating a procedure for selecting a background on the background selection mode of FIG. 32 .
  • the controller 110 controls to present the background selection screen in a background selection mode 830 as shown in a first screen image of FIG. 33 .
  • the background selection screen enables navigating background images using direction keys (navigation keys). If a selection key (OK key) is input while one of the background images is presented, the controller 110 returns to display the character house editing mode 820 and displays the selected background image as the background on the screen of the display unit 143 as shown in a second screen image of FIG. 33 .
  • the controller 110 presents available background images stored in the mobile phone with titles of the background images.
  • the background images are individually shown according to the navigation key manipulation. If a selection key is input while a background image is displayed, the background image is stored as the background and the character house editing screens are redisplayed.
  • FIG. 34 is a state diagram illustrating transitions between the character house editing mode and item editing modes of FIG. 25 .
  • FIGS. 35 a to 35 d illustrate screen images illustrating a procedure for adding an item to the character house in the character house editing mode of FIG. 34 .
  • the controller 110 displays a new item selection screen in a new item selection mode 860 as shown in FIG. 35 b.
  • the new item selection screen enables navigating item images using direction keys (navigation keys). Each item image is presented with a title of the item image.
  • the controller 110 displays an item positioning screen, as shown in FIG. 35 c, in an item positioning mode 870 .
  • the item positioning screen is partitioned to define a plurality of sections on the screen of the display unit 143 and each section is assigned a unique section number.
  • the item positioning screen facilitates the movement of an item from one place to another with reference to the section number.
  • the controller 110 displays a selected item editing screen, in which the selected item is positioned at the section having the input section number as shown in FIG. 35 d, in a selected item editing mode 850 .
  • the selected item editing screen facilitates fine adjustment of the selection item position by manipulating the navigation keys.
  • the selected item editing screen also provides an item-fixing key (for example, # key) for fixing the selected item at the adjusted position and a fast move key (for example * key) for quickly moving the item to another section in the item positioning mode 870 .
  • an item-fixing key for example, # key
  • a fast move key for example * key
  • the selected item editing mode 850 facilitates fine adjustment of the position of the selected item on the selected item editing screen.
  • the items can be overlapped each other.
  • the overlapped presentation of the items can be performed with a “send to front,” “send to back,” and “message adjustment options displayed when an option key is input. If the “send to front” (a “volume up” key) or the “send to back” (a “volume down” key), the controller 110 places the item on or beneath the item in the selected section.
  • a fast move key (* key) is selected while the mobile phone is in the selected item editing mode 850 , the mobile phone enters the item positioning mode 870 . If the item-fixing key is selected, the mobile phone enters an edited item shortcut selection mode 840 . If an “add item” option is selected, the mobile phone enters the new item selection mode 860 . If a “delete item” option is selected, the mobile phone enters the character house editing mode 820 . If a “background” option is selected, the mobile phone enters the background selection mode 830 .
  • the transition routes from the selected item editing mode are summarized as shown in table 5. TABLE 5 Functions in screen Fine position adjustment function (navigation key) having a selected Item overlapping display function (option >> item to be edited send to front, send to back, background)- background selection screen
  • FIG. 36 is a state diagram illustrating transitions between the character house editing mode and the selected item editing mode via the edited item shortcut selection mode of FIG. 25
  • FIGS. 37 a to 37 c show screen images illustrating a procedure for assigning shortcut key in the edited item shortcut selection mode of FIG. 36 .
  • an item to be edited is selected on the edited item shortcut selection screen in the edited item shortcut selection mode 840 .
  • the edited item shortcut selection mode 840 is transited by from the character house editing mode 820 represented by the character house editing screen as shown in FIG. 37 a and the selected item editing mode 850 represented by the selected item editing screen as shown in FIG. 37 b. If the character house editing screen is open, the controller 110 displays the items included in the character house as shown in FIG. 37 b.
  • a character house can be provided with maximal 10 items that are numbered from 0 to 9. Accordingly, if an item fixing key (#) is input while the character house editing screen is displayed, the controller 110 displays the edited item shortcut selection screen to assign the section numbers to the items.
  • the controller 110 displays the selected item editing screen to display the selected item in a bracket as shown in FIG. 37 c.
  • the selected item in the bracket can be finely adjusted in position. If the fast move key (*) is selected to quickly move the selected item at other section, the controller 110 displays the item positioning screen defining the sections.
  • the screen image presentation method for a mobile phone facilitates the design of a character house that is used as a screen image in a specific operation mode of the mobile phone.
  • the character house can be edited by adding, deleting, and changing positions of the background and items.
  • the screen image presentation method provides the mobile phone with the character house editing mode 820 in which the character houses stored in the mobile phone can be selected, deleted, and edited.
  • the mobile phone can switch between the character house-related editing modes for creating a new character house.
  • the background selection mode 830 is represented by the background selection screen which is accessed from the selected item editing screen of the selected item editing mode 850 or by selecting the “background” option in the menu provided in the selected item editing mode 850 .
  • the controller 110 presents the background image together with the title of the background image as shown in the first screen image of FIG. 33 .
  • the background image slides according to the direction key (left and right direction keys) manipulation to make the selection of one of the background images as the background of the character house possible.
  • the new item selection mode 860 facilitates the addition of a new item to the character house and transitions to the item positioning mode 870 and the selected item editing mode 850 .
  • the controller 110 displays the new item selection screen, as shown in FIG. 35 b, representing the new item selection mode 860 .
  • the controller 110 presents the item images together with the title of the items.
  • the new item selection screen facilitates the navigation of item images using direction keys (left and right direction keys). If one of the items is selected from the new item selection screen, the controller 110 displays the item positioning screen, as shown in FIG. 35 c, representing the item positioning mode 870 .
  • the item positioning screen 870 is partitioned so as to define several sections as shown in FIG.
  • the screen is divided into 9 sections and assigned unique section numbers, respectively. If a section number is selected while the item positioning screen, the controller 110 displays the selected item editing screen as shown in FIG. 35 d.
  • the selected item editing screen facilitates fine adjustment of the position of the selected item. If the addition of another item is required, the controller 110 displays the new item selection screen again, i.e. enters the new item selection mode 860 .
  • the controller 110 presents the items stored in the mobile phone if there is a command input for adding an item to the character house.
  • the screen is divided into several sections for quickly positioning an item on the screen and the item is finely adjusted in a section.
  • the controller 110 operates the edited item shortcut mode 840 and the selected item editing mode 850 to facilitate a change in the position of the item in the character house. That is, the controller 110 assigns the item numbers to the items presented with the character house on the edited item shortcut screen and highlights the item selected by the item number as a target item. The target item is edited in the selected item editing mode 850 and the item positioning mode 870 .
  • the controller 110 facilitates the editing of the character house at the steps 719 and 721 of FIG. 24 . In these steps, the controller 110 also determines whether to present the character quotients and the state and emotion of the character. Also, the controller 110 facilitates the setting of whether to activate the greeting behavior and random greeting event of the character.
  • the controller 110 also allows editing the background and items and adjusting the position of the items at steps 715 and 717 of FIG. 24 .
  • the character house includes the character, background, and items.
  • the character house can be created in the mobile phone and store at least one character house.
  • the controller 110 of the mobile phone facilitates the selection of one of the character houses in order for a selected character house to be used as a screen image in a specific operation mode of the mobile phone.
  • the background and items constituting the character house can be edited.
  • the background of the character house can be deleted, added, or modified in the character house editing mode under the control of the controller 110 .
  • the items of the character house can be deleted or added, and repositioned on the screen in the character house editing mode.
  • the controller 110 of the mobile phone performs the operations for presenting the character house at steps S 711 and S 713 of FIG. 24 .
  • the character house presentation is performed in the idle mode of the mobile phone at the same or a different time with a start of the idle mode. That is, the character house presentation can be set to start at the time when the display unit 143 is turned on by opening the clam shell phone or sliding up open the slide phone, or after completing an event of the mobile phone.
  • the character house presentation also can be set to start when a special event occurs. For example, the character house can be presented when verifying an incoming message, writing an outgoing message, or verifying a phone book.
  • the controller 110 controls such that the character house is presented in a transparency higher than that of the message or information, using the alpha blending method.
  • the character house is presented in 50% transparency while the message or information in 0% transparency.
  • the character house can be implemented with a display window such the message or information is presented in the display window of the character house.
  • the character house is preferably presented during a preset time to reduce power consumption of the mobile phone.
  • the controller 110 controls to present the character house (S 713 ).
  • the character house is presented while the controller 110 performs the steps S 313 to 315 of FIG. 14 . That is, the controller 110 verifies the character house configuration information (S 313 ) and determines whether the character house function is turned on (S 315 ). If the character house function is turned on, the controller 110 verifies the character house information (S 319 ) and then presents the character house in accordance with the character house information.
  • the character house includes a background, items, and character.
  • the controller 110 presents the background of the character house according to the procedure of FIG. 15 . If the background is a time-varying background, the controller 110 verifies the current time and presents a background image corresponding to the current time.
  • the controller 110 presents the items of the character house according to the procedure of FIG. 16 .
  • the controller 110 determines whether the character quotients presentation function is turned on. If the character quotients presentation function is turned on, the controller presents the character quotients. The controller 110 analyzes the types of items and presents the items according to the procedures corresponding to the types of the items. If the item is a motion item, the motion item is presented according to the procedure of FIG. 17 . If the item is a function item, the function item is presented according to the procedure of FIG. 18 . The presentation of the character is performed according to the procedure of FIG. 19 . If the greeting behavior function is set to “on,” the controller 110 verifies the current time and controls the character to show the greeting behavior if the current time is in a greeting time zone of a day.
  • the controller 110 controls to preset the greeting event according to the procedure of FIG. 21 .
  • the behavior of the character house is performed according to the procedure of FIG. 22 and a character movement is performed according to the procedure of FIG. 23 .
  • the controller 110 facilitates the setting of the character house functions, editing the character house, and presenting the character house according to the preset functions, according to the procedure of FIG. 24 .
  • the character house of the present invention is a virtual space in which an artificial intelligent character resides, talks, and moves such that a user recognize a state of a mobile by means of the character house.
  • the character house facilitates learning functions of the mobile phone and attracts the user's interest to the mobile phone.
  • the character house of an exemplary embodiment of the present invention is presented with a character, background image, items, and animation.
  • the character can present a sound effect and text expression reflecting its condition and emotion according to a character scenario, moves around the character house, and shows an animation.
  • the character can express the state and emotion of the mobile phone in various manners so as to make the character house dynamic.
  • the background is used to lively express the character house by changing the background image according to a progress of the time, whereby the user can guess the current time from the character house.
  • the items includes decoration items, motion items, and function items.
  • the function items show system information on the mobile phone to improve user convenience.
  • the decoration items increase the aesthetic sense of the character house.
  • the motion items move around in the character so as to make the character house dynamic.
  • the character house can be implemented with a special expression method in which the character greets the user of the mobile phone with an appropriate sound, text, and behavior at predetermined time zones of a day, to make the mobile phone user-friendly. Also, the character house can be implemented such that the character randomly gives a gift to the user, resulting in enjoyment of the user.

Abstract

An idle mode screen image presentation apparatus and method is provided for a mobile phone. A screen image presentation apparatus for a mobile phone according to the present invention includes a memory storing a character house including a background and function items for executing functions of the mobile phone; a controller presenting the character house in an idle mode of the mobile phone and providing information, when one of the function items is selected, on a function associated with the selected function item; and a display displaying the character house under the control of the controller.

Description

    CROSS-REFERENCE TO RELATED APPLICATION
  • This application claims the benefit under 35 U.S.C. §119(a) of Korean Patent Application filed in the Korean Intellectual Property Office on Dec. 6, 2005 and assigned Serial No. 2006-0118428, the entire disclosure of which is hereby incorporated by reference.
  • BACKGROUND OF THE INVENTION
  • 1. Field of the Invention
  • The present invention relates to portable phone. More particularly, the present invention relates to an idle mode screen image presentation apparatus and method for a mobile phone.
  • 2. Description of the Related Art
  • A portable terminal is typically a cordless electronic device such as a mobile phone. Recent mobile phones are implemented with various functions in addition to basic voice communication. Such phones include a camera phone equipped with a camera module, a digital TV phone equipped with a Digital Multimedia Broadcast (DMB) or Digital Video Broadcast (DVB) module, and an MP3 phone equipped with an MPEG-1 Audio Layer 3 (MP3) module. That is, the mobile phone has developed into a multipurpose electronic device.
  • Typically, an advanced mobile phone with an advanced multimedia processing ability is provided with a liquid crystal display (LCD) to display multimedia data. Various screen display techniques have been researched and developed to effectively display information. That is, a screen display scheme is designed to support viewing multimedia data while simultaneously displaying information related to a basic task. For example, a mobile phone can be implemented to display a preset image while the mobile phone is in an idle state. The image of the idle screen view is downloaded from a website or taken by a built in camera. However, conventional mobile phones limit designs to screen images to be presented on the display by themselves. Although there has been a screen image design method in which the screen image is designed with a background image, text, and clock selected by a user, most of the conventional screen image design techniques do not facilitate the design of a dynamic screen image with various character images.
  • Accordingly, there is a need for an improved screen image presentation apparatus and method for mobile phones.
  • SUMMARY OF THE INVENTION
  • An aspect of exemplary embodiments of the present invention is to address at least the above problems and/or disadvantages and to provide at least the advantages described below. Accordingly, an aspect of exemplary embodiments of the present invention is to provide a screen image presentation apparatus and method for a mobile phone.
  • It is another object of an exemplary embodiment of the present invention to provide a screen image presentation apparatus and method for a mobile phone that are capable of improving usability of mobile phone by using a character as an artificial intelligent agent.
  • According to another object of an exemplary embodiment of the present invention, a screen image presentation apparatus and method for a mobile phone are provided. This facilitates the presentation of a character house with an artificial intelligent character, items, and background as a screen image.
  • It is another object of an exemplary embodiment of the present invention to provide a screen image presentation apparatus and method for a mobile phone that are capable of making a presentation on a screen of the mobile phone dynamic by displaying an artificial intelligence-based character house, in which an intelligent character resides, as a screen image in a specific operation mode of the mobile phone.
  • It is yet another object of an exemplary embodiment of the present invention to provide a screen image presentation apparatus and method for a mobile phone that are capable of providing an intelligent character with an artificial intelligence-based character house. The intelligent character reacts and adapts to a change in the conditions of the mobile phone in the form of an animation, sound, and text.
  • It is another object of an exemplary embodiment of the present invention to provide a screen image presentation apparatus and method for a mobile phone that are capable of improving aesthetics of a screen image by means of a character house consisted of various visible items.
  • It is another object of an exemplary embodiment of the present invention to provide a screen image presentation apparatus and method for a mobile phone that are capable of informing progress of time by means of a character house having a background which is changed according to various time periods within a day.
  • According to another object of an exemplary embodiment of the present invention, a screen image presentation apparatus and method for a mobile phone is provided. This screen image presentation apparatus and method is capable of providing a character house as an animated screen image for a specific operation mode of a mobile phone. The character house includes a character, background, function items, decoration items and motion items. The character illustrates conditions and emotions reflecting states of the mobile phone in the form of an animation, sound, and text. The function items illustrate system information on the mobile phone. Decoration items improve aesthetics of the screen image and motion items move along respective movement lines in the screen house.
  • It is another object of an exemplary embodiment of the present invention to provide a screen image presentation apparatus and method for a mobile phone that are capable of producing current time zone information by means of a character house including a background and items and especially an intelligent character which illustrates greeting behaviors with an animation, sound, and text in the time zone.
  • It is yet another object of an exemplary embodiment of the present invention to provide a screen image presentation apparatus and method for a mobile phone that are capable of producing information on conditions and emotions representing states of a, mobile phone by means of character quotients, emoticons, and text expressions provided in a character house with a character, background, and items.
  • It is still another object of an exemplary embodiment of the present invention to provide a screen image presentation apparatus and method for a mobile phone that are capable of presenting a character house with a character, background, and items, as an idle mode screen image of the mobile phone.
  • It is another object of an exemplary embodiment of the present invention to provide a screen image presentation apparatus and method for a mobile phone that are capable of generating a character house with a character, background, and items, as a screen image for an idle mode of the mobile phone.
  • According to another object of an exemplary embodiment of the present invention, a screen image presentation apparatus and method for a mobile phone are provided. This apparatus and method facilitates a user's ability to design a character house, by selecting a character, background, and items, to be used as a screen image in an idle mode of the mobile phone.
  • It is another object of an exemplary embodiment of the present invention to provide a screen image presentation apparatus and method for a mobile phone that facilitates a user's ability to edit a character house consisting of a character, background, and items, by adding and deleting the items. The character house is used as a screen image for a specific operation mode of the mobile phone.
  • It is yet another object of an exemplary embodiment of the present invention to provide a screen image presentation apparatus and method for a mobile phone that are capable of adding new items or editing the items constituting a character house while the character house is presented as a screen image for a specific operation mode of the mobile phone.
  • It is still another object of an exemplary embodiment of the present invention to provide a screen image presentation apparatus and method for a mobile phone that are capable of adding new items to or removing or repositioning items presented with a character house. The character house consists of a background, character, and items and is used as a screen image for a specific operation mode of the mobile terminal.
  • According to another object of an exemplary embodiment of the present invention, a screen image presentation apparatus and method for a mobile phone is provided. The screen image presentation apparatus and method facilitates a user's ability to select a partitioned sectors of a character house to see the selected sector in an enlarged form such that the user can perform a fine adjustment of a position of an item located in the selected sector. The character house consists of a background, character, and items, and is used for a screen image for a specific operation mode of the mobile terminal.
  • In accordance with an aspect of an exemplary embodiment of the present invention, the above and other objects are accomplished by a screen image presentation apparatus of a mobile phone. The screen image presentation apparatus includes a memory, a controller and a display. The memory stores a character house including a character for presenting a condition of the mobile phone. The controller controls to present the character house at a predetermined presentation time and the display displays the character house under the control of the controller.
  • In accordance with another aspect of an exemplary embodiment of the present invention, the above and other objects are accomplished by a screen image presentation method for a mobile phone. According to the screen image presentation method, configuration information on a character house is verified with a character that represents a condition which is determined by usage pattern of the mobile phone. The character house is presented in a specific operation state of the mobile phone on the basis of the configuration information.
  • In accordance with another aspect of an exemplary embodiment of the present invention, the above and other objects are accomplished by a screen image presentation method for a mobile phone. According to the screen image presentation method, configuration information of a character house is registered in a program mode. The character house with a background and items is edited in an editing mode. The configuration information of the character house is checked and the character house is presented with a character, items, and background on the basis of the configuration information of the character house.
  • Other objects, advantages and salient features of the invention will become apparent to those skilled in the art from the following detailed description, which, taken in conjunction with the annexed drawings, discloses exemplary embodiments of the invention.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The above and other exemplary objects, features and advantages of certain exemplary embodiments of the present invention will be more apparent from the following description taken in conjunction with the accompanying drawings, in which:
  • FIG. 1 is a block diagram illustrating a configuration of a mobile phone employing a screen image presentation apparatus and method according to an exemplary embodiment of the present invention;
  • FIGS. 2 a, 2 b and 2 c are views illustrating a concept of a character house of a screen image presentation apparatus according to an exemplary embodiment of the present invention;
  • FIG. 3 is a screen image illustrating an exemplary character house according to an exemplary embodiment of the present invention;
  • FIGS. 4 a to 4 n are screen images illustrating steps for implementing exemplary character animations in a screen image presentation technique according to an exemplary embodiment of the present invention;
  • FIGS. 5 a to 5 c are screen images illustrating steps for implementing an animated background in a screen image presentation technique according to an exemplary embodiment of the present invention,
  • FIG. 6 illustrates screen images that illustrate background images for implementing an animated background in a screen image presentation technique according to an exemplary embodiment of the present invention;
  • FIG. 7 a is screen image illustrating a character quotient module on the idle mode screen in a screen image presentation technique according to an exemplary embodiment of the present invention;
  • FIG. 7 b is an enlarged view illustrating the character quotient module of FIG. 7 a;
  • FIG. 8 is a drawing illustrating a table of emoticons to be displayed with the quotient module of FIG. 7 b.
  • FIGS. 9 a to 9 l are views illustrating the presentation of function items in a screen image presentation technique according to an exemplary embodiment of the present invention;
  • FIGS. 10 a to 10 c are views illustrating steps of presenting motion items in a screen image presentation technique according to an exemplary embodiment of the present invention;
  • FIGS. 11 a to 11 c are views illustrating steps of presenting decoration items in a screen image presentation technique according to an exemplary embodiment of the present invention;
  • FIG. 12 illustrates screen images illustrating character behaviors in the screen image presentation technique according to an exemplary embodiment of the present invention;
  • FIG. 13 is a screen image illustrating a character animation with an event in a screen image presentation technique according to an exemplary embodiment of the present invention;
  • FIG. 14 is a flowchart illustrating a character house presentation procedure of a screen image presentation method according to an exemplary embodiment of the present invention;
  • FIG. 15 is a flowchart illustrating the background presentation procedure of FIG. 14;
  • FIG. 16 is a flowchart illustrating the item presentation procedure of FIG. 14;
  • FIG. 17 is a flowchart illustrating a procedure for presenting a motion item in the screen image presentation technique according to an exemplary embodiment of the present invention;
  • FIGS. 18 a to 18 d are flowcharts illustrating a procedure for presenting function items in the screen image presentation technique according to an exemplary embodiment of the present invention;
  • FIG. 19 is a flowchart illustrating a character presentation procedure of the screen image presentation method according to an exemplary embodiment of the present invention;
  • FIG. 20 is a flowchart illustrating a greeting behavior presentation procedure of FIG. 19;
  • FIG. 21 is a flowchart illustrating a greeting event presentation procedure that occurred during a greeting behavior of the character of the screen image presentation technique according to an exemplary embodiment of the present invention;
  • FIGS. 22 a and 22 b are flowcharts illustrating an animation behavior presentation procedure of the screen image presentation technique according to an exemplary embodiment of the present invention;
  • FIGS. 23 a and 23 b illustrate the character shift information determination procedure in detail;
  • FIG. 24 is a flowchart illustrating a screen house presentation procedure of a screen image presentation method for a mobile phone according to another exemplary embodiment of the present invention;
  • FIG. 25 is a state diagram illustrating a character house design procedure of a screen image presentation method according to an exemplary embodiment of the present invention;
  • FIG. 26 is a state diagram illustrating transitions between the character house configuration mode and the character house editing mode of FIG. 25;
  • FIG. 27 is a screen image illustrating a character house configuration screen in the character house configuration mode of FIG. 25;
  • FIG. 28 is a state diagram illustrating transitions between the character house configuration mode and the character house editing mode of FIG. 25;
  • FIG. 29 illustrates screen images illustrating a procedure for setting a “character house presentation on/off” option on the character house configuration screen of FIG. 25;
  • FIG. 30 is a state diagram illustrating transitions among the character house editing mode and specific item editing modes of FIG. 25;
  • FIG. 31 illustrates a screen image illustrating an example character house editing screen in the character house editing mode of FIG. 25;
  • FIG. 32 is a state diagram illustrating transitions between the character house editing mode and the background selection mode of FIG. 25;
  • FIG. 33 illustrates screen images illustrating a procedure for selecting a background on the background selection mode of FIG. 32;
  • FIG. 34 is a state diagram illustrating transitions between the character house editing mode and item editing modes of FIG. 25;
  • FIGS. 35 a to 35 d illustrate screen images illustrating a procedure for adding an item to the character house in the character house editing mode of FIG. 34;
  • FIG. 36 is a state diagram illustrating transitions between the character house editing mode and the selected item editing mode via the edited item shortcut selection mode of FIG. 25;
  • FIGS. 37 a to 37 c illustrate screen images illustrating a procedure for assigning shortcut key in the edited item shortcut selection mode of FIG. 36.
  • Throughout the drawings, the same drawing reference numerals will be understood to refer to the same elements, features and structures.
  • DETAILED DESCRIPTION OF EXEMPLARY EMBODIMENTS
  • The matters defined in the description such as a detailed construction and elements are provided to assist in a comprehensive understanding of the embodiments of the invention. Accordingly, those of ordinary skill in the art will recognize that various changes and modifications of the embodiments described herein can be made without departing from the scope and spirit of the invention. Also, descriptions of well-known functions and constructions are omitted for clarity and conciseness.
  • A “screen image” is an image presented on a display screen of the mobile phone.
  • “House,” “avatar house,” and “character house” are used interchangeably to refer to a visible object presented on screen in a specific mode of the mobile phone. The house is displayed with characters, items, and background.
  • “Item” is a visible object for enriching the house and includes decoration items, function items for notifying states of specific functions, and motion items that move along preset paths.
  • “Background” is an image forming a background graphics of the house.
  • “Character” is a visible object that resides in the character house and can express a state of the mobile phone in an emotional manner. The character includes fictional characters represented by a human, avatar, or animal, among others that can express various emotions. Accordingly, the character can express emotions.
  • In the exemplary embodiment of the present invention, the character house can be correlated with a background, a character, and at least one item.
  • FIG. 1 is a block diagram illustrating a configuration of a mobile phone employing a screen image presentation apparatus and method according to an exemplary embodiment of the present invention.
  • Referring to FIG. 1, the mobile phone includes a controller 110, data processing unit 120, radio frequency (RF) unit 123, audio processing unit 125, voice synthesizer 127, memory 130, keypad unit 141, and display unit 143.
  • The RF unit 123 facilitates communication between the mobile phone and other terminals over a wireless channel. The RF unit 123 includes an RF transmitter for up-converting and amplifying the frequency of a transmitted signal, and an RF receiver for low-noise amplifying a received signal and down-converting the frequency of the received signal.
  • The data processing unit 120 includes a transmitter for coding and modulating the transmitted signal, and a receiver for demodulating and decoding the received signal. The data processing unit 120 may include a modem and a codec. The voice synthesizer 129 performs a function of converting a text message output from the controller 110 or received from outside into a voice signal.
  • The keypad unit 141 is provided with a plurality of alphanumeric keys for input data and functions key for setting various functions.
  • The memory 130 includes a program memory and a data memory. The program memory stores programs for processing incoming and outgoing calls, and programs for presenting the character house for a specific state of the mobile phone and controlling reactions of the characters in response to events caused by operations of the mobile phone and for structuring the screen image comprising the character house with the characters, according to an exemplary embodiment of the present invention. The data memory includes a character memory for storing character images and information on the characters for performing the exemplary embodiment of the present invention and a text memory for storing text data for emotional expressions of the characters. The data memory also temporarily stores data generated while the programs are executed. The term “character house image” is interchangeably used to refer to the screen image presented in a specific mode of the mobile phone. The specific mode or specific state may refer to the idle mode or idle state.
  • The character memory may include a character storage, an item storage and a background storage. The character storage stores information on the structures of the characters and images thereof. The item storage stores information on the structures of the items and image thereof. The background storage stores background images. The character house structure may be stored in the form of information on a combination of the character, item, and background image, and the images stored in the respective storages may be used individually.
  • A character is provided with information on character animations expressing emotional states for the mobile phone, character images for expressing emotions in response to text messages (for example, SMS, EMS, and MMS, among others), and character images for expressing emotions in response to events occurring in the mobile phone. According to an exemplary embodiment of the present invention, a plurality of character images expressing the emotional states for the mobile phone, a plurality of character images expressing the emotions in accordance with the contents of the text messages, and a plurality of character images expressing the emotions in accordance with the events occurring in the mobile phone is assumed. The memory for storing the images operates in a file structure. A character file structure includes a header and character images, and the header contains information on identifiers, start delimiters, a number of images, display positions and names of animations. In a character transmission mode, the headers and character images are transmitted after being processed in the form of a transmission file.
  • The character house is displayed on the screen together with a background image, a character, and at least one item in a specific mode (for example, an idle mode) of the mobile phone. That is, the character house is designed by arranging the items on the background image for a specific mode screen (for example, an idle mode screen), the items and the background images being selected by the user. At this time, the file structure of the character house may include a header and a plurality of images, and the header may include a header of the character house and the item headers of the items associated with the character house. The character house header contains information on the character house identifier, the positions of the characters on the screen, the sizes of the characters, and the names of the characters; and each item header contains information on the item index, the position of the item in relation with the character house, and the name of the item. In a character house transmission mode, the headers of the character house and the items associated with the character house and the background and item images are transmitted in the form of a compressed file.
  • The items means the images presented in association with the character house on the screen. Each item may be displayed in a specific association with the character house or regardless of the character house. If the item is not associated with the character house, the item should be transmitted individually. The item includes a header and an image, and the item header contains information on the item index, the position of the item in relation to the character house, and the name of the item. According to an exemplary implementation, if the item is not associated with the character house, the information on the position of the item has a default value.
  • In this exemplary embodiment of the present invention, the character is used for presenting the information on the states and event occurrences of the mobile phone. The character house is used to indicate the specific operation mode (for example, the idle mode) of the mobile phone. The item is used to show the functions of the mobile phone, decorate the character house, and show the screen dynamic. The background image is used to provide a background of the characters and items associated with the character house.
  • The controller 110 controls general operations of the mobile phone. The controller 110 may be integrated with the data processing unit 120. According to an exemplary implementation, the controller 110, data processing module 120, audio processing unit 125, and voice synthesizer 127 can be integrated into a mobile station modem (MSM) chip. In this exemplary embodiment of the present invention, the controller 110 controls to select the characters, items, background image, character house, texts, and sounds. The controller arranges the selected images and texts at the present positions on the screen of the display unit 143 and outputs the texts as voice sounds under the support of the voice synthesizer 127.
  • The display unit 143 displays the state of the mobile phone under the control of the controller 110. The display unit 143 may be implemented with a liquid crystal display (LCD) together with a LCD controller, extra memory for storing the display data, and LCD display device. If the display unit 143 is implemented with a touchscreen function, the display unit 143 can operate an input device.
  • The characters represent a graphical user interface (GUI) including the character displayed on the screen of the display unit 143 in the specific mode of the mobile phone. The character house is a space in which the character with an artificial intelligence resides such that the user can experience a virtual space with a virtual talking character. The user can quickly learn the functions of the mobile phone with the assistances of the character house and enjoy with the vivid characters.
  • The character house has a function that expresses a dynamic character. First, the character house can express the condition and emotion of the current character in the animation, sound, and text according to a character scenario associated with the character. Second, the character house has a function to present a background of the screen for a specific mode (especially, the idle mode) of the mobile phone that is changed in accordance with a progress of time. Third, the character house has a function for displaying items. The items displayed in association with the character house may be classified by function. In this exemplary embodiment of the present invention, the items can be classified into three categories. The first is the function item for presenting the functions and system information on the mobile phone in order to enhance the user's convenience. The second is the decoration item for increasing aesthetic satisfaction by adding various images to the character house. The third is the motion items that make the character house screen dynamic by adding the motion images (for example, the cloud or bird, among others).
  • The character house may further provide a greeting function with which the character associated with the character house gives greetings with appropriate sounds at preset times. An additional function such as allowing a user to give a present to the character may also be implemented.
  • FIGS. 2 a to 2 c are views illustrating a concept of a character house of a screen image presentation apparatus according to an exemplary embodiment of the present invention.
  • FIG. 2 a is an illustration of a house setting screen for selecting a character house to be displayed on the screen. The house setting screen is displayed on the screen under the control of a character setting program. If the character setting program is executed, 4 character house images are displayed in corresponding tiled windows on the screen of the display unit 143. When a number of the character house images are less than 4, the windows that are not allocated for the character house image will be empty.
  • If a character house is selected from the house setting screen, the controller 110 operates a character engine which is structured as illustrated in FIG. 2 b. The character engine stores information of the selected character house into a character house information storage module 210 and executes a character house expression module 220 upon activation. That is, the character engine includes the character house information storage module 210 and the character house expression module 220. The character house information storage module 210 is a module for storing information on the character houses and the character to be displayed in association with each other. According to an exemplary embodiment of the present invention, the character house information storage module 210 includes character house information storage 213, system information storage 215, and character information storage 217.
  • The character information storage 217 stores emotion information, status information, quotient information, and text and sound information associated with the character's emotions and statuses. The character information storage 217 also stores animation information for animating the character in a specific status. The mobile phone processes various events which have occurred while operating. The mobile phone also updates the quotients of the characters that correspond to the events.
  • The quotient information includes a physical quotient (HQ), emotion quotient (EQ), intelligent quotient (IQ), and social quotient (SQ), that are used for determining the status and emotion of the character. The controller 110 accumulates the quotients associated with the events whenever an event is processed and determines the status of the mobile phone on the basis of the quotient analysis results. The emotion of the character is determined according to the quotients and user's feedback analysis results. The emotion of the character can be determined in accordance with a content related to the event (for example, a text message) that is obtained during the process. The text and sound information associated with the emotion can be stored together with the character expression. The character information storage 217 includes information on the animation of the character that is expressed in a specific mode (for example, an idle mode) of the mobile phone.
  • The system information storage 215 stores system information on the function items for displaying a memory capacity of the mobile phone, scheduled alarms and unread mail. The items associated with the character house includes the decoration items, motion items, and function items. The function items present the operations and functions of the mobile phone. According to an exemplary embodiment of the present invention, the items for presenting a memory status, preset alarm times, and unread mail are described as examples of the function items.
  • The character house information storage 213 stores the information on the character house that is defined by a character house setting program such that the character house information includes the types and positions of the items and the type of the background associated with the character house. The character house information storage 213 stores the information on the character house images except for the character and the function items. Thus, the character house information 213 stores the information on the decoration items, motion items and the background images.
  • The character house expression module 220 includes a character display module 223, an item display module 225 and a background display module 227.
  • The character display module 223 renders the character animation of the character house. The character display includes a mobility and entrance of the character and a text balloon or a sound effect associated with the behavior of the character. The character display module 223 can make the character express special motions such as a greeting. The item display module 225 displays the items at preset positions with reference to the character house information on the current character house and the system information. The items are displayed in a different manner by item category (decoration, motion, and function). The background display module 227 displays the background image in accordance with the information stored in the character house information storage 213. The background image can be configured such that the brightness of the background image changes over time.
  • FIG. 2 c illustrates an exemplary character house screen displayed by the character engine of FIG. 2 b. The character house is displayed in a specific mode (for example, an idle mode) of the mobile phone and includes a character, items and a background.
  • FIG. 3 is a screen image illustrating an exemplary character house according to an exemplary embodiment of the present invention. The character house screen provides information on the detailed status, which is not provided with a normal idle mode screen, for the mobile phone together with decoration effects, as illustrated in FIG. 3.
  • Referring to FIG. 3, a character 230 is displayed in the form of an animated character in association with the character house. The character 230 behaves in a motion preset for the idle mode of the mobile phone to alarm at preset times in corresponding animation together with a text and sound effect. The character 230 can be set to animate for giving items as presents for the character house in a random manner. In order to express the state of the mobile phone and the emotion of the character representing the state of the mobile phone, the character 230 expresses the emotion in the animation according to the state of the mobile phone and an emotion model together with the sound and text. The background 240 is represented by a background image of the character house and the brightness of the background can be changed over time.
  • The items 250 give effects for decorating the character house and inform the operations of the mobile phone. The function items 251, 253, 255, and 259 decorate the character house and present the statuses of the preset functions and operations. For example, the function item 251 is a clock item that illustrates a preset alarm time, and the function item 253 is a mailbox item for illustrating the amount of unread mail. If multiple alarms are set, the closest alarm time is displayed. The function item 255 is a memory monitor item for displaying the state of the memory. The decoration item 259 is an item for decorating the character house. The motion item (not shown) can be activated for decorating the character house.
  • The character quotient module 260 includes a character quotient, a physical quotient, an intelligent quotient (IQ) and an emotional quotient (EQ). In an exemplary embodiment of the present invention, the social quotient (SQ) is represented by values assigned for the events related to the communication. The physical quotient (HQ) is represented by values assigned for events related to the performance for the mobile phone. The intelligent quotient (IQ) is represented by values assigned for an amount of information stored in the mobile phone and the emotional quotient (EQ) is represented by values assigned for events related to the multimedia. The character quotient module 260 can be configured to be fixed at a position or floats across the screen.
  • The character 230 of the character house has the following three basic expression factors: animation, text and speech. Animation facilitates the character's 230 ability to express a state and an emotion while moving. A text facilitates the character's 230 expression of a state and emotion in text. The text can be presented in the form of a text balloon. Speech facilitates the character's 230 expression of a state and emotion through voice or sound effects. The character 230 can express various states and emotions, and the state and emotion of the character 230 are determined in accordance with a state decision model algorithm and an emotion mode decision algorithm as a character scenario dependent on the variation of the quotient values of the character. The decision on the state and emotion mode of the character on the basis of the character scenario is disclosed in the Korean Patent Application No. 2005-107174. The character 230 of the character house expresses the state information and the emotion mode using the character image together with a unique voice effect and text balloon.
  • The character 230 can be implemented with an animation in which the character 230 moves around the character house and disappears outside the screen and then appears from outside. That is, the character 230 expresses the state and emotion with a motion, sound effect, and text, if the time to express the state and emotion has arrived. The state and emotion of the character 230 is expressed when the quotients reach respective threshold values, rather than always. However, the character house is always displayed in the specific mode screen. For example, if the mobile phone enters an idle mode, the character house is displayed on the screen. The character house screen disappears when a preset display time expires or when the mobile phone enters another operation mode. If the character house is displayed at the idle mode, the character 230 expresses the state and emotion in motion and with other animations.
  • FIGS. 4 a to 4 h are screen images illustrating steps for implementing an exemplary character animation in a screen image presentation technique according to an exemplary embodiment of the present invention. FIGS. 4 i to 4 n are screen images illustrating steps of another exemplary character animation in a screen image presentation technique according to an exemplary embodiment of the present invention.
  • Referring to FIGS. 4 a to 4 h, the character 230 moves in a preset behavior pattern in the character house. The character 230 appears at a random position on the screen, makes a speech expression and then moves around. If the character house disappears on the screen while the character moves around the character house, the character 230 appears at a random position and repeats the behavior after a predetermined time has elapsed. The motion, sound effect, and text can be selectively disabled. The character house can be expressed through six steps as followings.
  • As illustrated in FIG. 4 a, if the mobile phone enters the idle mode, the items and background image of the character house are displayed. Next, as illustrated in FIG. 4 b, a character 230 appears at a position in the character house. The character 230 can appear with a fade-in effect. That is, the character 230 can be implemented with an alpha blending technique so that the transparency of the character 230 changes from 100% to 0% gradually. For example, the transparency reduces by 25% per 0.5 second such that the character 230 is gradually faded in to completely appear in 2 seconds. The character 230 starts appearing at a random position. The character 230 is preferably implemented so that the character is at a position in which the character does not completely appear. That is, the character 230 is configured to appear at a random position but where the character 230 can be completely shown across the border of the screen.
  • After the character 230 is displayed, a sound effect is output in accordance with the state of the character and with a notification icon, as illustrated in FIG. 4 c. The sound effect is output at once and repeated one more time when the character reappears. The sound effect may also be cancelled.
  • After the sound effect and the text are output, the character 230 is animated as illustrated in FIG. 4 d. The character animation process is repeated until the character 230 disappears. The character animation is performed at a fixed position.
  • As illustrated in FIG. 4 e, the state and emotion of the character 230 is expressed with a text after the character animation starts. The text is expressed in the text balloon in consideration of the time taken for the user to read the text. The text balloon is displayed in the form of a semitransparent image with about 30% of transparency and is open in about 3 seconds in consideration of the reading time.
  • After the text balloon disappears, the character 230 moves as illustrated in FIGS. 4 f and 4 g. The movement direction of the character 230 can be changed according to the position at which the character 230 appears. That is, the position at which the character 230 appears is verified such that the character 230 moves in a direction to which the character can secure the longest distance for movement. The character 230 moves at a predetermined speed. The character 230 is preferably set to move at the speed of about 20 pixels per second. The character 230 disappears across the border of the screen. If the character 230 disappears completely, an initial idle mode screen is displayed as illustrated in FIG. 4 h identical with FIG. 4 a. After a predetermined time (about 2 seconds) elapses, steps 4 a to 4 g are repeated.
  • The character animation illustrated in FIGS. 4 i to 4 h is an example in which the character does not move. Referring to FIGS. 4 i to 4 h, the character is displayed in the idle mode screen. In the initial state of the idle mode illustrated in FIG. 4 i, the preset items and background images are displayed without the character. Next, as shown in FIG. 4 j, the character appears on the screen. The character appears at a predetermined position on the screen. The appearance position is changed to the bottom right position in relation to the center of the screen. The appearance position can be changed according to the size of the character. The character can be set to appear with a fade-in effect in such a manner that the transparency of the character changes from 100% to 0% gradually. For example, the transparency reduces by 25% per 0.5 second such that the character is gradually faded in to completely appear in 2 seconds.
  • After the character 230 is completely displayed, a sound effect is output in accordance with the state of the character and with a notification icon as illustrated in FIG. 4 k. The sound effect is output at once and repeated one more time when the character reappears.
  • According to an exemplary implementation, a text is displayed to express the current state and emotion of the character. The text is presented in the text balloon with a transparency of 30%. The text can be displayed in synchronization with a sound. The text balloon is expressed in consideration of the time taken for the user to read the text. For example, the text balloon can be set to disappear after 1.5 seconds.
  • After the sound effect ends and the text balloon disappears, the character is animated as illustrated in FIG. 4 m. The character animation process is repeated until the character disappears in the character house as illustrated in FIG. 4 n.
  • The character animation process is performed in a certain order so that the items 250 and background image 240 are displayed, the character 230 appears in a fade-in effect, the sound effect is output for notifying the state of the mobile phone, the text is presented for notifying the state of the mobile phone, and the character animation is performed in the character house.
  • The character animation is a motion behavior of the character without moving around the screen, and the character movement is a behavior of the character moving around the screen. The character animation and the movement may be performed simultaneously. The character 230 may also appear simultaneously with the items 250 and the background. The character 230 may appear in the form of an opaque object, such as with a 0% transparency. The output of sound effect and text may be skipped.
  • According to an exemplary implementation, the presentation of the background can be performed so that the background image changes according to the progress of the time. A fixed background image can be used.
  • FIGS. 5 a to 5 c are screen images illustrating steps for implementing an animated background in a screen image presentation technique according to an exemplary embodiment of the present invention. FIG. 6 illustrates screen images illustrating background images for implementing an animated background in a screen image presentation technique according to an exemplary embodiment of the present invention. The background images of FIG. 6 can be used for the animated background or the fixed background.
  • Referring to FIGS. 5 a and 5 c, the animated background is implemented with several background images that are presented in series over a period of time. According to an exemplary embodiment of the present invention, a day is divided into time periods, such as morning, afternoon, evening, and night. The background is placed with different background images for the respective time periods. In this exemplary embodiment of the present invention, the background images of the character house change according the specific time period if an animation background 240 is set by the user. Since the change of the background is performed in accordance with the progress of the time, the background of the character house that changed with different brightness background images is assigned for the morning, day, and night as illustrated in FIGS. 5 a to 5 c. According to an exemplary embodiment of the present invention, the time periods are defined as illustrated in table 1. According to an exemplary implementation, the character can be set to present corresponding greetings at the beginning of each time period.
    TABLE 1
    Class Time Period Description
    Morning 06:00˜12:00 A little Bright Image having sunshine
    Afternoon 12:00˜18:00 Warm Image having Sun in the middle sky
    Evening, 18:00˜06:00 Dark image with no sun
    Night
  • The background may be set on the character house setting screen. FIG. 6 illustrates the background images that may be selected for the background. The background images may be used for the animated background in which background images are changed in accordance with the time period or may be used for a fixed background. The background images can be used for the animated background of the character house or can be used for the fixed background of the character house. Each background image can be used for the animated background by only changing the brightness of the background image for expressing the time periods.
  • The character house can present information on items that illustrate the system information and on the quotients of the character, as well as the character 230 and the background 240.
  • FIG. 7 a is screen image illustrating a character quotient module on the idle mode screen in a screen image presentation technique according to an exemplary embodiment of the present invention. FIG. 7 b is an enlarged view illustrating the character quotient module of FIG. 7 a.
  • Referring to FIG. 7 a, the character quotient module 260 notifies the current status of the character 230. In this exemplary embodiment of the present invention, the character quotient module 260 illustrates the state and emotion of the character 230 with the HQ, EQ, IQ, and SQ status gauges. Since the state and emotion of the character are presented with the quotient status gauges, the character can attract the user's interest.
  • The character quotient module 260 can be presented at a preset position in the character house. In this exemplary embodiment of the present invention, the character quotient module is presented at the top left area of the screen, as illustrated in FIG. 7 a. The character quotient module 260 can be set to be visible or invisible according to a user selection.
  • Referring to FIG. 7 b, the character quotient module 260 includes a quotient information display region 267, an emotion icon display region 263, and a status display region 265. The emotion icon display region 263 can be implemented to display emoticons representing the emotions of the character. The status display region 265 presents the status of the character in text and the quotient information display region 267 displays the levels of the quotients with color bars (gauge). The emotions that can be presented in the emotion icon display region 263 includes “happy,” “normal,” “angry,” “sad,” and “tired,” and the emotions are linked to corresponding images in the form of a table as illustrated in FIG. 8. Accordingly, when implementing the character house, the emotion of the character can be presented by the emoticon of the character quotient module corresponding to the current emotion of the character in addition with the character animations. The status information of the character can be presented at the status display region 265 in the form of a text. Each character can have various statuses that can be expressed in the form of a text. The character quotient information is displayed in the form of gauges in the quotient information display region 267. In this exemplary embodiment of the present invention, four quotients, such as, HQ, EQ, IQ, and SQ are measured and presented in the quotient information display region 267. In order to distinguish the quotients from each other, the quotients are presented by means of colored gauges, such as a red gauge for SQ, a green gauge for HQ, a purple gauge for IQ, a blue gauge for EQ, as shown in FIG. 7 b. Each gauge presents the level of the quotient in a percentage (%) without a text based notification. The emotion icon display region 263 and a status display region 265 can be deleted.
  • The items and expressions such as the greetings can be bought in an item shop on the web or received as a gift. The items that can be presented in the character house include the function items, decoration items, and motion items. The positions of the items on the screen can be set in the character house setting screen. Presentation of the items is explained based on an order of the function item, motion item, and decoration item.
  • The function items are items that provide specific system information of the mobile phone. In this exemplary embodiment of the present invention, the function items include a memory monitor item, an alarm item, and a mailbox item. Additional function items may also be added when required. The mailbox item can be a message item, which is used to explain the function item.
  • The function items can be implemented for a signal function or can be an item grouping the function items comprising similar functions. As a group-wise function item, the mailbox item can present a variety of the messages. According to an exemplary implementation, the function item integrates a short message service (SMS) item, an email message service (EMS) item, and a multimedia message service (MMS) item. When there is an alarm item, a schedule item, a memo item, and an alarm generation item, among others may be integrated. If such a multifunction item is selected, the controller 110 can display the function items integrated in the multifunction item such that the user can select one of the items.
  • FIGS. 9 a to 9 l are views illustrating how to present function items in a screen image presentation technique according to an exemplary embodiment of the present invention.
  • The functional item can be displayed in a normal mode (see FIG. 9 a), a selection mode (see FIG. 9 b), and an enable mode (see FIG. 9 c). Presentation of the function item with a single function is described with reference to FIGS. 9 a to 9 c.
  • Referring to FIG. 9 a, when the function items are presented with the character house, the function items are arranged at the preset positions on the screen. In FIG. 9 a, the mailbox item, the clock item, and the memory monitor item are examples displayed on the screen. In the normal mode, the function items are displayed as basic item images. If an event occurs while the function items are displayed in the normal mode, the basic item image of the function item related to the event is animated. In the normal mode, the mobile phone enters a selection mode for displaying the function item information if an item display mode key (a special key such as “clear key” provided to the mobile phone) is input.
  • Referring to FIG. 9 b, the function items are displayed with a semitransparent layer in the selection mode, and the selected item is zoomed in to be displayed in an enlarged form. In the selection mode, other function items can be selected by keypad manipulation. If an OK key is input while an item is selected, the item is activated in the enable mode.
  • Referring to FIG. 9 c, upon entering the enable mode, the item image of the selected function item is displayed together with the information on the selected function item. By manipulating the up/down direction keys provided in the keypad unit, it is possible to navigate to other applets or return to the normal mode shown in FIG. 9 a. Left/right direction keys allow navigating to and activating other function items.
  • A structure and presentation of a function item are described with the memory monitor item as an example hereinafter with reference to FIGS. 9 d to 9 f.
  • The memory monitor item is a function item for display memory utilization information. FIG. 9 d illustrates a memory monitor item image in the normal mode. If the memory usage is over 50% in the normal mode, the memory monitor item is animated as illustrated in FIG. 9 e. If the memory monitor item is selected by the keypad manipulation, the memory monitor item is displayed on the screen in the form of an enlarged memory monitor item image presenting the memory utilization with a text and gauge. The memory utilization information is displayed in the text form of a used memory/total memory. Also, the memory utilization information is presented with a colored bar of the gauge. The gauge bar can be placed in left or right side of the memory monitor item according the position of the memory monitor item on the screen and changed in color (for example, red) if the memory utilization percentage is over a predetermined percentage (for example, 95%). If a “go to memory status” command is input through a menu screen, a detailed memory utilization screen is displayed to provide information on the usage of the memory.
  • A structure and presentation of a function item are described with the clock item as an example hereinafter with reference to FIGS. 9 g to 9 i.
  • The clock item is a function item for alarming at preset times. FIG. 9 g illustrates the clock item image in the normal mode. If a preset time has arrived, the clock item is animated as illustrated in FIG. 9 h. If the clock item is selected by the keypad manipulation, the clock item is presented on the screen in the form of an enlarged clock item image with the present time. The preset time can be presented in the form of a text and image. The preset time can be of a variety of functions such as “morning call,” “schedule,” “to do,”. The text can be presented in various formats. For example, a text format 1 can be used for a title of the schedule and a text format 2 for date and time. If the schedule title is not set, a default text “no title” is presented. In the text format, the time is presented in the form of TT/MM, and the date is presented in the form of today/tomorrow/a day of the week. The text can be positioned at a left or right side of the clock item according to the position of the clock item on the screen. If multiple schedules are preset, the most closely arranged schedule is shown. If a “go to alarm status” command is input through a menu screen, a detailed schedule screen is displayed for informing preset alarm schedule details.
  • A structure and presentation of a function item are described with the mailbox item as an example hereinafter with reference to FIGS. 9 j to 9 l.
  • The mailbox item is a function item for informing the number of the unread messages among the received messages. FIG. 9 j illustrates the mailbox item image in the normal mode. If a message is received in the normal mode, the mailbox item is animated as illustrated in FIG. 9 k. If the mailbox item is selected by the keypad manipulation, the mailbox item is presented on the screen in the form of an enlarged mailbox image with the information on the received and unread messages. The information on the received and unread messages can be presented with a text and image. The text can be presented
  • The text can be presented in various formats. For example, a text format 1 can be used for a category of the information (in this case, New Message), and a text format 2 for the numbers of the messages in the form of “new message/total message.” The text can be positioned at a left or right side of the mailbox item according to the position of the mailbox item on the screen. If a “go to inbox” command is input through a menu screen, a message list screen is displayed for illustrating the received messages.
  • The motion items are items for making the screen presentation of the character house dynamic. The motion items move around on the screen according to movement lines by item.
  • FIGS. 10 a to 10 c are views illustrating steps of presenting motion items in a screen image presentation technique according to an exemplary embodiment of the present invention.
  • Referring to FIGS. 10 a to 10 c, the motion items of the character house include animals and movable objects as shown in FIG. 10 a. Each motion item moves in accordance with a unique movement line and moves at a preset animation time.
  • The motion items are animated on the basis of the position information contained in the character house information such that each motion item starts moving at the preset time along the preset movement line as illustrated in FIG. 10 b. The motion item can be configured to move at a random speed and turn the progress direction in random probability.
  • As illustrated in FIG. 10 c, when moving across each other, the motion items are overlapped. According to an exemplary implementation, the motion items are illustrated in an order of the presentation priorities (depth) that are preset for the respective items. All the motion items are assigned presentation priorities so that the motion item with the highest priority is presented when at least two motion items are overlapped while on the move. Some motion items have higher priorities than that of the character, and others have lower priorities. In FIG. 10 c, the motion item “collie” (a dog) has a priority higher than the character such that the legs of the character are covered by the collie. Also, a special motion item (in FIG. 10 a, a butterfly) is assigned the highest priority such that it always appears on the character and other items.
  • As described above, the motion items are animated to move at a certain speed (for example, 5 pixels/sec) so they may disappear out of the character house (cross the border of the screen) and then re-appear. If a motion item encounters the character or other item while moving along the movement line, the motion item or the character with a higher priority is displayed on top. If a motion item with a priority lower than that of the character passes across the character, the motion item is shaded by the character.
  • The decoration items are objects for decorating the character house such that the decoration items have no special functions. FIGS. 11 a to 11 c are views illustrating steps in the presentation of decoration items in a screen image presentation technique according to an exemplary embodiment of the present invention. FIG. 11 a illustrates exemplary decoration items that can be presented for decorating the character house, FIG. 11 b illustrates exemplary character houses decorated with the decoration items, and FIG. 11 c illustrates specific decoration items shown in an enlarged form on the screen.
  • Referring to FIGS. 11 a to 11 c, the decoration items to be applied to the character house include flowers, furniture, and other ornamental things. Such decoration items can be presented with the character in the character house as illustrated in FIG. 11 b. Some decoration items have special meanings. The special decoration items such as a trophy and a clover shown in FIG. 11 c are obtained as a reward for winning a game or for a certain achievement of a specific state of the mobile phone, but bought in cash. The decoration item “trophy” is used for informing a number of wins of the game, such that all wins of games may change the trophy in color and type and increases the number of the trophies.
  • Until now, how to present the character 230, the background 240, the items, and the character quotient module 260 is explained. Here, the character quotient module 260 can be selectively set to be visible or invisible. All the elements constituting the character house are presented in association with the character house. The character house also can be set to differently operate at a specific time or situation. For example, the character house can be configured such that the character gives greetings at specific times or when a specific event occurs.
  • When the greeting behavior is set, the character gives the user greetings at the preset times and an item or character money can be given as a reward for the behavior.
  • FIG. 12 shows screen images illustrating character behaviors in the screen image presentation technique according to an exemplary embodiment of the present invention. As shown in FIG. 12, when the character greets at a specific time, the greeting expression is displayed in the form of a text balloon and the greeting expression changes according to the time periods of the day.
  • The greeting behavior of the character can be configured such that the character greets at a predetermined number of times per day (such as, 4 times a day). If a determination is made that the character greets 4 times a day, 4 greeting time zones are configured: morning (08:30˜09:30), afternoon (12:00˜13:30), evening (17:30˜18:30), and night (23:30˜00:30). Accordingly, the character greets once at each greeting time zone. If the mobile phone is off or in character house disabled modes at each greeting time zone, the greeting behavior is disabled. Therefore, the mobile phone is not in the idle mode where the character house screen is presented (for example, when a clam shell type phone is unfolded or a slide type phone is opened), the greeting behavior function is disabled. When the greeting behavior starts, the greeting behavior remains during a predetermined time period (for example, 10 seconds). If a specific event (for example, an incoming call or a key input for activating a specific function) occurs while the character greets, the greeting is performed if the greeting behavior is maintained over a predetermined amount of time (for example, 3 seconds). If the greeting behavior stops before 3 seconds elapse and the character house is activated at the greeting times zone, the greeting behavior starts again. The character is preferably associated with different greeting behaviors for the respective greeting time zones as in table 2.
    TABLE 2
    Class Time Zone Text
    Morning 08:30˜09:30 Hello!
    Afternoon 12:30˜13:30 Good˜
    Evening 17:30˜18:30 Bye˜
    Night 23:30˜00:30 See U˜
  • When the greeting behavior starts, the mobile phone can generate an event. The event may occur in a probability set according to the times of the greeting behaviors. In this exemplary embodiment of the present invention, the probability is about 10%. If the event does not occur during the greeting time zone (that is, the mobile phone is off during the greeting time zone or the character house is not activated), the event will become invalid.
  • FIG. 13 is a screen image illustrating a character animation with an event in a screen image presentation technique according to an exemplary embodiment of the present invention.
  • Referring to FIG. 13, if the character animation starts, a text is presented in the form of a text balloon. The text expression as an even is presented in the text balloon. The text balloon is displayed with a 30% transparency. The text balloon for expressing the event can be presented together with an image. According to an exemplary embodiment of the present invention, a gift event text is presented with a gift image as shown in FIG. 13. The gift image is preferably presented at a bottom left or bottom light portion of the screen. The gift event can be replaced by money or other item event.
  • The items and background are selected and arranged on the screen for decorating the character house.
  • FIG. 14 is a flowchart illustrating a character house presentation procedure of a screen image presentation method according to an exemplary embodiment of the present invention.
  • Referring to FIG. 14, the controller 110 controls the operations of the mobile phone and determines whether the mobile phone enters a character house presentation mode (S311). If the mobile phone enters the character house presentation mode, the controller verifies the character house configuration information (S313). The character house presentation mode can be identical with the idle mode of the mobile phone. In the idle mode, the controller 110 displays the character house on the screen of the display unit 143 during a predetermined display period. If the display period ends, the controller 110 turns off the display unit 143 to save power. If a clam shell or a slide of the mobile phone is open in the idle mode, the controller 110 turns on the display unit 143. The character house configuration information includes parameters configured using a setting application. The parameters may include information for activating or deactivating the character house function. The character house configuration information may include character house on/off information, character information display on/off information. If the character house configuration information includes the character house function related information, the controller 110 determines whether the character house function is turned on (S315). If it is determined that the character house function is turned off, the controller 110 displays a normal idle mode screen.
  • If it is determined that the character house function is turned on, the controller 110 verifies the character house information (S319). The character house information is on the currently set character house. The character house information includes the information on the background and items associated with the character house. The background information includes a background image ID, and the item information includes item IDs and positions of the items on the screen (X, Y coordinate). The character house can be displayed with one or more items, and the items are arranged in an order of the presentation priorities such that if two items are overlapped, the item with a higher priority is displayed on the other item. By verifying the character house information, the controller 110 recognizes the background and items with their positions on the screen of the display unit 143.
  • After verifying the character house information, the controller 110 controls to present the background, items, and character (S321, S323, and S325). In FIG. 14, the character house presentation is performed in the order of the background, items, and character. However, the presentation order of the elements can be changed. For example, the elements of the character house can be presented in an order of the background, character, and items; an order of the character, background, and items; and an order of the character, items, and background. That is, the character house presentation can be performed in various orders. Through the steps S321 to S325, the character house presentation is completed.
  • During the presentation of the character house, the mobile phone determines whether an event (for example, a function is selected by a key input) occurs or a predetermined display time ends (S327). If an event occurs or the display time ends, the controller 110 ends the presentation of the character house and then performs the selected function or turns off the display unit. Also, it is possible to perform the selected function while maintaining the character house screen. For example, if a message is received while the character house is displayed or a message transmission mode is activated, the controller 110 can maintain the presentation of the character while reading the received message and writing a message to transmit. The controller 110 displays a message received or a message to be transmitted in a transparent message window on the character house screen. Also, the controller 110 can increase the transparency of the character house presentation and display the message on the character house presentation. Other events can be processed while the character house is presented in the similar manner.
  • FIG. 15 is a flowchart illustrating the background presentation procedure of FIG. 14. Referring to FIG. 15, the controller 110 verifies a background ID of the character house (S361), and determines whether the background ID is 0 (S363). The background ID “0” means the character house does not use any background images. If the background ID is 0, the controller 110 loads a normal background image set as a default image (S365). If the default background image is not set, no background image is presented.
  • If the background ID is not 0, the controller 110 retrieves a background corresponding to the background ID and presents the retrieved background on the screen. According to an exemplary implementation, the controller 110 selects one of the backgrounds as shown in FIG. 6 on the basis of the background ID.
  • The background of the character house can be represented by multiple images that are presented at different time periods. The background of the character house can be represented by a fixed single image.
  • Accordingly, after retrieving the background, the controller 110 determines whether the retrieved background is a multiple image background (S364). If the retrieved background is a multiple image background, the controller verifies the current time (S367) and determines that the current time is greater than or equal to 06:00 and less that 12:00 (S369). If the current time is greater than or equal to 06:00 and less than 12:00, the controller 110 selects a morning image assigned for the morning or afternoon from the multiple image background and then displays the morning image on the screen (S371). If the current time is not greater than or equal to 06:00 or if the current time us not less than 12:00, the controller 110 determines whether the current time is greater than or equal to 12:00 and less than 18:00 (S373). If the current time is greater than or equal to 12:00 and less than 18:00, the controller 110 selects an afternoon image assigned for the afternoon from the multiple image background and then displays the afternoon image on the screen (S375). If the current time is not greater than or equal to 12:00 or not less than 18:00, the controller 110 determines whether the current time is greater than or equal to 18:00 and less than 23:59, or greater than or equal to 00:00 or less than 06:00 (S377). If the current time is greater than or equal to 18:00 and less than 23:59, or greater than or equal to 00:00 or less than 06:00, the controller 110 selects an evening or night image assigned for the evening or night from the multiple image background and displays the evening or night image on the screen (S379). At step S364, the retrieved background is not a multiple image background, the controller regards the retrieved image as a fixed single image background and displays the fixed signal image as the background (S381).
  • When the multiple image background is used, a signal fixed image can be used by changing the brightness of the image for the respective time periods of a day.
  • FIG. 16 is a flowchart illustrating the item presentation procedure of FIG. 14. The item presentation procedure of FIG. 16 is described with the character information presentation of FIG. 7 b. However, the character information presentation can be performed with the background presentation procedure, or the character presentation procedure of FIG. 19.
  • In this exemplary embodiment of the present invention, it is assumed that the character information presentation is included in the item presentation procedure as shown in FIG. 16. The emoticons for expressing the emotion of the character, text for expressing the state of the character, and the character quotients, shown in FIG. 7 b, included in the character information are explained.
  • Referring to FIG. 16, the controller 110 verifies whether character house configuration information includes information on a character quotient presentation function (S401) and determines whether the character information presentation function is on (S403). If the character information presentation function is on, the controller 110 verifies the character quotient information (S405) and presents a character emoticon on the basis of the character quotient information in the emotion icon display region 263 of FIG. 7 b (S407). The emotion of the character can be represented by 5 emotion modes: happy, normal, angry, sad, and tired. The emoticon to be displayed in the emotion icon display region 263 can be one of the emoticons corresponding to the emotion modes, as shown in FIG. 8. Once the emoticon is presented, the controller 110 presents a text corresponding to the current state of the character in the status display region 265 of FIG. 7 b (S409) and then presents the character quotients (SQ, HQ, IQ, and EQ) in the quotient information display region 267 of FIG. 7 b in the form of bar graphs (S411). The values of character quotients are presented as different colored bars. As mentioned above, if the character information presentation function is enabled, the controller 110 verifies the character quotients and determines the emotion mode on the basis of the character quotients. If the emotion mode is determined, the controller 110 presents an emoticon corresponding to the emotion mode, a text appropriate for the state of the character, and the quotients in the form of the bar graphs.
  • The character information based on the character quotients can be presented at a fixed position in the character house (for example, a top left portion) as shown in FIG. 7 a in the form of FIG. 7 b. The character quotient information cannot be presented according to the user's configuration. The user decides whether to present the character information. If the user does not activate the character information presentation function, the controller recognizes the configuration at step S403 such that the steps S405 to S411 are skipped. According to an exemplary implementation, the character information based on the character quotients information shown in FIG. 7 b is not presented.
  • After presenting the character quotients, the controller verifies the information on an item with a lowest priority (depth) (S413). The item information includes an item ID and display position. The item can be one of the motion items, function items, and decoration items. In some cases, two or more items can be overlapped. The items are presented in an order of the presentation priorities assigned to the items such that the item with a lowest priority is presented first. That is, when two or more items are overlapped, only the item with the highest priority is completely presented. If the item on top is disabled, the item with the second highest priority is presented.
  • Accordingly, after verifying the item information, the controller 110 determines whether the item information exists (S415). If the item information exists, the controller 110 determines whether the item is a motion item on the basis of the item information (S417). If the item is a motion item, the controller 110 presents the motion item (S419) and verifies another item with the next lowest priority (S421). If it is determined that the item is not a motion item at step S417, the controller 110 determines whether the item is a function item (S423). If a determination is made that the item is a function item, the controller 110 presents the function item (S425). If the item is not a function item, the controller 110 presents the item as a decoration item (S427).
  • The item presentation is performed in a different manner based on the type of item. That is, the motion item, function item, and decoration item are presented in different procedures.
  • FIG. 17 is a flowchart illustrating a procedure for presenting a motion item in the screen image presentation technique according to an exemplary embodiment of the present invention.
  • Referring to FIG. 17, the controller 110 verifies the motion item information on the basis of the item ID (SS431). The motion item information includes an item ID, a sojourn time for staying the item at the initial position, a coordinate represented by a horizontal distance (X axis distance) for moving the item to the next position and a vertical distance (Y axis distance) for moving the item to the next position, and initial position information. The initial position information contains one of 8 coordinates listed in table 3. The motion item is assumed to start moving from an edge of the screen as the initial position (such as 4 sides and 4 corners of the screen of the display unit 143) in table 3. The motion item may also be configured to start moving at a random position in the screen of the display unit 143.
    TABLE 3
    1. (X=0)
    2. (X=LCD_X_MAX)
    3. (Y=0)
    4. (Y=LCD_Y_MAX)
    5. (X=0, Y=0)
    6. (X=0, Y=LCD_Y_MAX)
    7. (X=LCD_X_MAX, Y=0)
    8. (X=LCD_X_MAX, Y=LCD_Y_MAX)
  • After verifying the motion item information, the controller 110 obtains the initial position of the motion item to animate the motion item at the initial position (S433). The animation is maintained before the sojourn time is ended and then the motion item is positioned at the next position through the steps S435 to S445. After animating the motion item, the controller 110 determines whether the sojourn time expires (S435). If the sojourn time is expired, the controller 110 calculates a coordinate of the next position to which the motion item moves (S437). After calculating the next position coordinate, the controller 110 determines whether the calculated next position is in a available display range on the screen of the display unit 143 (S439). If the calculated next position is in the available display range, the controller 110 moves the motion item to the next position (S441) and then returns to step S433 to animate the motion item at the next position. At step S439 the controller 110 determines that the movement of the motion item is completed and moves positions the motion item at the initial position (S443) and repeats the step S433 if the calculated next position is not in the available display range.
  • The motion item is configured with one of the initial positions described in table 3 and moves to a predetermined coordinate on the screen after a predetermined period of time. The motion item is animated at the initial position, moved to the next position after the sojourn time, and then animated again, under the control of the controller 110. The movement time is determined on the basis of the sojourn time and a distance between the two positions. If the sojourn time is short and the distance between the two positions is close, the motion item moves fast. Alternatively, if the sojourn time is long and the distance between the two positions is great, the motion item moves slowly. The movement direction of the motion item depends on the relationship of the two positions (coordinate in the screen). The motion item can move in horizontal, vertical and diagonal directions and can be determined in consideration of the distance between the two positions. The motion item can be one of the items depicted in FIG. 10 and move in the manner illustrated in FIG. 10 b. The motion items move on the screen as illustrated in FIG. 10 b in accordance with the procedure of FIG. 17.
  • The movement of the motion item stops once the character house presentation ends. The movement of the motion item may also be set to stop at a time different from the character house presentation end time. That is, the motion item presentation time can be shorter than the character house presentation time. According to an exemplary implementation, the motion item disappears before the character house presentation stops.
  • Accordingly, if the sojourn time is not expired at step S435, the controller 110 maintains the animation of the motion item (S445) and determines whether the motion item presentation time is ended, (S447). If the motion presentation time is ended, the controller stops presentation of the motion, and otherwise, repeats the step S435.
  • FIGS. 18 a to 18 d are flowcharts illustrating a procedure for presenting function items in the screen image presentation technique according to an exemplary embodiment of the present invention.
  • Referring to FIGS. 18 a and 18 b, the controller 110 verifies a type of function item (S452). In this exemplary embodiment of the present invention, three function items are assumed to exist. This items are the mailbox item, alarm clock item, and memory monitor item. The number and type of the function items can be changed.
  • After the function item is verified, the controller 110 determines whether the function item is a mailbox item (S453). If the function item is the mailbox item, the controller verifies a presentation position of the mailbox item and presents mailbox item at the presentation position on the screen (S455). Next, the controller 110 verifies a number of received but unread messages (S457) and determines whether the number of unread messages are greater than 0 (S459). If the number of the unread messages is not greater than 0, the controller presents an empty mailbox image.
  • If the number of the unread messages is greater than 0, the controller 110 displays the number of the unread messages with the mailbox image (S461 to S467). The number of the unread messages can be presented in various manners. For example, the number of unread messages can be displayed on the mailbox image. Also, the number of unread messages can be displayed by means of a text balloon. According to an exemplary implementation, the text balloon has a shape of a postcard with the number of the unread messages. In this exemplary embodiment of the present invention, the mailbox item presentation is explained with the text balloon image for displaying the number of the unread messages.
  • According to the position of the mailbox item, the controller 110 determines an unread message number presentation position. The controller 110 verifies the left and right side spaces of the mailbox items and arranges the text balloon at the left or right side. If a determination is made that the text balloon is arranged at the right side, the controller 110 determines whether the text balloon presentation position is in the available display range of the screen (S463). If the text balloon presentation position is in the available display range, the controller 110 presents the text balloon at the right side of the mail box (S467). If the text balloon presentation position is not in the available display range, the controller 110 determines to the text balloon at the left side of the mailbox item (S465) and repeats the step S463.
  • At step S453, if it is determined that the function item is not the mail box item, the controller determines whether the function item is an alarm clock item (S469) in FIG. 18 b. If the function item is the alarm clock item, the controller 110 verifies scheduled alarms on the basis of alarm information (S471). The alarm information includes the presentation position of the alarm clock item on the screen of the display unit 143. After verifying the scheduled alarms, the controller determines whether the there is scheduled alarms (S473). If no alarm is scheduled, the controller 110 presents an empty alarm image. If there is at least one scheduled alarm, the controller 110 retrieves a next scheduled alarm and displays an alarm clock image with the next scheduled alarm (S475). The alarm clock item can be implemented with a clock image. According to an exemplary implementation, the alarm clock image may have the needles arranged to show the scheduled alarm time.
  • Next, the controller 110 verifies an alarm time presentation position and determines whether there is enough space to present the alarm time at the right side of the alarm clock item in the screen. If there is enough space to present the alarm time, the controller 110 presents the next scheduled alarm time at the right side of the alarm clock item (S477). If it is determined that there is not enough space to present the alarm time at the right side of the alarm clock item (S479), the controller 110 presents the alarm time at the left side of the alarm clock item (483). The alarm can be represented by a morning call, a meeting schedule, and a reservation, among others. The scheduled alarms are activated at the scheduled times with the specific events. The alarm time can be shown in the form of YYYY/MM and HH/MM as a text. The alarm time text can be placed at the left or right side of the alarm clock item according to the position of the alarm clock item.
  • If two or more alarms are scheduled, the alarm time of the next closest alarm is presented. The alarm time can be presented by means of the alarm clock image. When the alarm time is presented by means of the alarm clock image. The alarm time text may not be presented.
  • If it is determined that the function item is not the alarm clock item at step S469, the controller 110 determines that the function is a memory item. Next, the controller 110 verifies a presentation position of the memory item from the character house information and presents the memory item at a verified presentation position (S484). After presenting the memory item on the screen, the controller 110 verifies a total memory capacity and used memory capacity (S485) and then displays the information on the total memory capacity and the user memory capacity (S487 to S493). That is, after verifying the memory information, the controller 110 verifies the side of the memory item that has enough space for presenting the memory information. If it is determined that the right side of the memory item has enough space for presenting the memory information, the controller 110 presents the memory information at the right side of the memory item. Alternatively, if it is determined that the right side of the memory item does not have enough space for presenting the memory information, the controller 110 presents the memory information at the left side of the memory item. The memory item may inform of the memory information of the memory unit 130 of FIG. 1 and the memory information can be shown in the form of a gauge bar. The gauge bar can be located at the left or right side according to the position of the memory item on the screen. The gauge bar shows the memory utilization and changes in red if the memory utilization reaches over 90%.
  • As describe above, a function item is presented in such a manner that the controller verifies the presentation position of the function and presents the function item at the presentation position on the screen of the display unit 143 together with the function information obtained by the function item. The function information is presented in such a manner that the controller verifies which side of the function item has enough space for presenting the function information and then presents the function information at a left or right side of the function item, where there is enough space. The function information can be shown as a text or a graphical gauge. When the function information is provided in the text, the graphical gauge may or may not be used.
  • FIGS. 18 c and 18 d are flowcharts illustrating a procedure for presenting function items in the screen image presentation technique according to another exemplary embodiment of the present invention.
  • Referring to FIGS. 18 c and 18 d, the controller 110 verifies a type of function item (S751). In this exemplary embodiment of the present invention, three function items are assumed to exist. The function items include a mailbox item, alarm clock item, and memory monitor item. The number and types of the function items can be changed. If function information is associated with a function item, the controller presents the function item in the form of an animated function item for attracting the user's interest and presents the function information in a popup window during a popup display time. In this exemplary embodiment of the present invention, the function items are presented in the following order: mailbox item, alarm clock item, and memory item.
  • After the function item is verified, the controller 110 determines whether the function item is the mailbox item (S753). If the function item is the mailbox item, the controller 110 verifies an amount of the unread mail (S755) and then determines whether the amount of the unread mail is greater than 0 (S757). If the amount of the unread mail is 0, the controller 110 presents an empty mailbox image on the screen of the display unit 143 as shown in FIG. 9 j (S767) to notify that there is no unread message.
  • If the amount of the unread mail is greater than 0, the controller 110 presents an animated mailbox image as shown in FIG. 9 k. After presenting the animated mailbox image, the controller determines whether a popup time is arrived (S761). If the popup time has arrived, the controller 110 presents a popup window as shown in FIG. 9 l (S763). The popup window is displayed during a predetermined popup time and then disappears. During the presentation of the popup window, the controller 110 simultaneously starts a popup counter. After the popup window is presented, the controller determines whether the popup counter is expired (S765). If the popup counter is expired, the popup window disappears.
  • If the function item is not the mailbox item as in step S753, the controller 110 determines whether the function item is an alarm clock item (S771). If the function item is the alarm clock item, the controller 110 verifies scheduled alarms (S773) and determines whether there are scheduled alarms (S775). If there is no scheduled alarm, the controller 110 presents an empty alarm image on the screen of the display unit 143 for notifying that no scheduled alarm exists as shown in FIG. 9 g (S785). If there is at least one scheduled alarm, the controller 110 presents an animated alarm clock image as shown in FIG. 9 h (S777) and determines a popup time has arrived (S779). If the popup time is arrived, the controller 110 presents a popup window as shown in FIG. 9 i (S781). The popup window is displayed during a predetermined popup time and then disappears. During the presentation of the popup window, the controller 110 simultaneously starts a popup counter. After the popup window is presented, the controller 110 determines whether the popup counter is expired (S783). If the popup counter is expired, the popup window disappears. The popup window presentation time may be configured in consideration of the alarm time. According to an exemplary embodiment of the present invention, the scheduled alarm is assumed to be an alarm set with a specific date and time for “meeting” and “to do” rather than an everyday occurrence alarm (for example, morning call), and displays alarms before 6 hours.
  • If the function item is not the alarm clock item, the controller 110 determines whether the function item is a memory monitor item (S786). If the function item is the memory monitor item, the controller 110 verifies memory utilization (S787) and determines whether the memory utilization is in a predetermined normal range (S789). If the memory utilization is in the normal range, the controller 110 presents only the memory monitor item image on the screen of the display unit 143 for notifying that the memory utilization is normal, as shown in FIG. 9 d (S799). If the memory utilization exceeds the normal range, the controller 110 presents an animated memory monitor as shown in FIG. 9 e (S791) and determines a popup time has arrived (S793). If the popup time is arrived, the controller 110 presents a popup window as shown in FIG. 9 f (S795). The popup window is displayed during a predetermined popup time and then disappears. During the presentation of the popup window, the controller 110 simultaneously starts a popup counter. After the popup window is presented, the controller 110 determines whether the popup counter is expired (S797). If the popup counter is expired, the popup window disappears. The memory monitor item can be displayed regardless of the memory utilization. According to an exemplary implementation, the controller 110 verifies the memory utilization and then presents the memory utilization in the form of a popup window.
  • As described above, the function items can be presented together with the character house. According to an exemplary implementation, the function items are displayed in a preset order. While presenting the function items in the preset order, the controller determines whether to present the function item in an active state. If it is required to present the function item in the activate state, the controller 110 presents the function item as an activated function item image and then provides the information related to the function in the form of a popup window after a while. The popup window is presented during a predetermined popup time. For example, when all the function items are presented in the activate state, the controller 110 presents in the order of the mailbox item as an animated mailbox image followed by the mailbox popup window, the alarm clock item as an animated alarm clock image followed by the alarm popup window, and the memory monitor item as an animated memory monitor image followed by the memory utilization popup window. The function items are presents in the character house as shown in FIGS. 9 j, 9 g, and 9 d after the popup windows disappear.
  • The decoration items are presented on the basis of the decoration item information contained in the character house information. The item information includes an item ID that provides information on the type of the decoration item and a presentation position on the screen. The controller 110 retrieves a decoration item image corresponding to the item ID and presents the retrieved decoration item image at the presentation position on the screen as shown in FIG. 11 a. The decoration item is used for improving aesthetic factor of the character house, regardless of any function. The decoration item can be added or removed according to the user's configuration.
  • In the character house, the character is implemented with three behavior properties.
  • First, the character greets the user of the mobile phone at preset times. Second, the character irregularly gives the user a gift while greeting the user. Third, the character expresses a state of the mobile phone or an emotion reflecting the state of the mobile phone in an animation or a motion. In this exemplary embodiment of the present invention, it is assumed that the mobile phone has various emotions that represent a user's contribution to the mobile phone. The character house can be presented only with the third behavior property. That is, the greeting and gift-giving behavior properties can be disabled when the character house is presented. In this exemplary embodiment of the present invention, it is assumed that the greeting, gift-giving, and animation behavior properties are enabled.
  • The character house can consist of the items and background as well as the character, and the character house is presented in a specific operation mode of the mobile phone. When the character house is presented in the idle mode, the character house can be shown as an idle mode screen (for example, when a clam shell of the mobile terminal is unfolded or a slide mobile phone is slid open). If an event occurs while the character house is presented, the character house disappears. In other words, the idle screen is replaced by other function screen. Also, the character house can be maintained when a specific event occurs (for example, a message processing event).
  • FIG. 19 is a flowchart illustrating a character presentation procedure of the screen image presentation method according to an exemplary embodiment of the present invention.
  • Referring to FIG. 19, the controller 110 verifies a current time (S501) and then determines whether the current time is in any of preset greeting behavior time zones (S503, S513, S517, and S521). In this exemplary embodiment of the present invention, the greeting behavior time zones are set to four time zones as in the table 2. The greeting behavior is shown once in a greeting behavior time zone.
  • Accordingly, the controller 110 determines whether the current time zone is in the greeting behavior time zones (S503, S513, S517, and S521). If the current time zone is in the greeting behavior time zones, the controller 110 determines the greeting behavior is shown (S505, S515, S519, and S523).
  • If it is determined that the current time zone is not in the greeting behavior time zones at the steps of S503, S513, S517, and S521 or greeting behavior is performed already in the greeting behavior time zone at the steps S505, S515, S519, and S523, the controller 110 skips the greeting behavior and controls the character to be in a normal behavior (S511).
  • If it is determined that the current time zone is in the greeting behavior time zones at the steps of S503, S513, S517, and S521 and the greeting behavior is not performed yet in the greeting behavior time zones at the steps S505, S515, S519, and S523, the controller 110 records the current time (S507) and controls the character to show the greeting behavior (S509). Recording the current time zone helps to avoid situations in which the greeting behavior is not performed in the same greeting behavior time zone. After the greeting behavior is performed, the controller 110 controls the character to be in the normal behavior (S511).
  • The greeting behavior presentation procedure is performed as shown in FIG. 20. The greeting behavior is presented as the character greets to the user and can be performed with a salutation and an addition event for giving an item or character money in a probability.
  • FIG. 20 is a flowchart illustrating a greeting behavior presentation procedure of FIG. 19.
  • Referring to FIG. 20, the controller 110 verifies a current greeting time zone (S531) and verifies current character information (S533). Next, the controller 110 controls the character to show the greeting behavior in the form of an animation (S535) and retrieves a greeting text corresponding to the current greeting time zone (S537). After retrieving the greeting text, the controller 110 presents the retrieved greeting text in the form of a text balloon (S539). The greeting behavior can be presented as shown in FIG. 12. The greeting text can be presented in the form of the text balloon of FIG. 12. The greeting text can be changed, depending on the greeting time zones.
  • When the character house presentation is terminated due to an event that occurs during the character's greeting behavior, the controller 110 determines that the greeting behavior has been performed if the greeting behavior is performed for a tolerable time. According to an exemplary implementation, the tolerable time is assumed to be 3 seconds.
  • Accordingly, the controller 110 determines whether the greeting behavior is continued over the tolerable time (in this embodiment, 3 seconds) (S541). If the greeting behavior is not continued over the tolerable time, the controller 110 maintains the greeting behavior and text balloon presentation (S557). While maintaining the greeting behavior and the text balloon presentation, the controller determines an event occurs such that the character house presentation is terminated in 3 second-tolerable time (S559). If an event occurs in tolerable time, the controller 110 stops the greeting behavior and then returns. According to an exemplary implementation, the greeting behavior is performed again, if the character house is activate in the same greeting behavior time zone.
  • If the greeting behavior is performed over the tolerable time (3 seconds), the controller determines that the greeting behavior is completely performed and records the completion of the greeting behavior in the greeting behavior time zone (S543). Next, the controller 110 determines whether the event presentation time has arrived (S545). The event may occur in a random probability set according to the times of the greeting behaviors.
  • If the event presentation time has arrived, the controller 110 presents the greeting event (S547).
  • FIG. 21 is a flowchart illustrating a greeting event presentation procedure that occurred during a greeting behavior of the character of the screen image presentation technique according to an exemplary embodiment of the present invention.
  • The greeting event can randomly occur while the character performs the greeting behavior. The greeting event may also be presented in the same manner with the character's greeting behavior except for the screen layout and the text.
  • Referring to FIG. 21, the controller 110 selects an item or character money as a gift at the greeting event presentation time (S561) and presents an image corresponding to the item or the character money to give as the gift (S563). Next, the controller 110 verifies the information of the character (S565) and animates the character (S567). Subsequently, the controller 110 presents the greeting text in the form of the text balloon (S569).
  • When presenting the character according to the greeting event, the controller 110 animates the character with the text and/or sound as shown in FIG. 13. At this time, the emotion mode of the character is fixed to “happy” and the character is presented with the item image and/or the character money image as the gift. The item image and the character money image provided in the greeting event presentation procedure can be displayed at the bottom left or bottom right portion of the character house as shown in FIG. 13, and the text expression related to the gift is presented in the form of a text balloon having a 30% transparency.
  • After performing the greeting event presentation procedure as in FIG. 21 at step S547 of FIG. 20, the controller 110 determines whether 10 seconds are elapsed after the presentation of the greeting expression text balloon (S549). If 10 seconds have elapsed, the controller 110 makes the text balloon and the character animation image disappear (S551).
  • If 10 seconds have not elapsed, the controller maintains the presentation of the current character animation and the greeting expression text balloon (S553). The greeting behavior of the character is assumed to be maintained for 10 seconds. Accordingly, before 10 seconds have elapsed during the presentation of the greeting behavior of the character, the controller 110 maintains the greeting behavior of the character (S549, S553, and S555). If 10 seconds have elapsed for the start of the greeting behavior of the character, the controller 10 makes the greeting expression text balloon and the character animation image disappear (S551).
  • As described above, if the current time is in a predetermined greeting time zone, the controller 110 controls the character of the character house to show the greeting behavior corresponding to the greeting time zone. The greeting behavior is for greeting to the user of the mobile phone and is performed on in every greeting time zone. If the mobile phone is turned off or the character house is not activated, the greeting behavior of the character is skipped. In this exemplary embodiment of the present invention, the greeting behavior of the character is performed in a random probability to give the user an unexpected gift. In this exemplary embodiment of the present invention, the gift-giving behavior is performed during the greeting behavior time. However, the gift-giving behavior may be performed at any time during the presentation of the character house. If the gift-giving behavior is performed at any time during the character house presentation, the controller 110 can control the gift-giving behavior to be performed when the character is presented.
  • In FIG. 19, the controller 110 controls the behavior of the character so that it is normal (S511) after performing the greeting behavior of the character out of the greeting time zones or after performing the greeting behavior. The character behaviors are presented in the procedure of FIG. 22 a or FIG. 22 b.
  • FIGS. 22 a and 22 b are flowcharts illustrating an animation behavior presentation procedure of the screen image presentation technique according to an exemplary embodiment of the present invention.
  • In FIG. 22 a, the character is moving and animated together with a text and sound for expressing the state and emotion of the character while the character house is presented. In FIG. 22 b, the character is animated after the text and sound expression is presented in the character house.
  • The character behavior can be vividly expressed in the character house in various manners. For example, the character can be represented by a plurality of animation images such that the animation images are continuously displayed to form an animation of the character. Also, the character can be presented to move along a preset movement line. The character animation is implemented by presenting the animation images in series at regular intervals, similar to presenting the function items as shown in FIG. 9 e, 9 h, and 9 k. The character movement is implemented by moving the character on the basis of a preset movement direction, movement distance, and movement time. In order for the character to seamlessly move, the movement can be implemented with a plurality of images. FIG. 23 shows a procedure for presenting a character movement.
  • FIG. 22 a shows a character movement behavior presentation procedure, and FIG. 22 b shows a character gesture behavior presentation procedure.
  • Referring to FIG. 22 a, the controller 110 verifies information on the current character (S571) and determines an initial appearance position of the character (an X, Y coordinate in the screen of the display unit 143). The initial appearance position of the character can be randomly determined in a display range of the screen of the display unit 143. After the initial appearance position is determined, the controller 110 presents the character with 100% transparency (S575). The 100% transparency produces character invisibility so that the background and items images are displayed on the screen of the display unit 143 as shown in FIG. 4 a.
  • According to an exemplary implementation, the controller 110 decreases the transparency of the character image at a predetermined rate so that the character gradually appears in the character house. The controller 110 controls such that the character appears gradually by repeating the steps S577 and S579. If the transparency of the character image is assumed to decrease by 10%, the controller 110 controls to gradually reduce the transparency of the character image to be opaque (alpha blending effect). Accordingly, the transparency of the character image decreases from 100% to 0% by 10% such that the character appears in the character as shown in FIG. 4 b.
  • According to an exemplary implementation, the controller 110 outputs a sound effect representing the state of the character as shown in FIG. 4 c (S581), animates the character as shown in FIG. 4 d (S583), presents a text expressing the state and emotion of the character in the text balloon during a predetermined presentation time as shown if FIG. 4 e, and then terminates the animation of the character. According to an exemplary implementation, the presentation time is set to 3 seconds. During the presentation of the character house as discussed above, the text balloon is presented after appearance of the character in the order of FIGS. 4 a to 4 e and presentation of the sound effect and the character animation.
  • The controller 110 determines a next position of the character and then performs movement of the character (S589, S597, and S559). In order to present the character movement, the controller determines position shift information (S587). The position shift information is determined according to a procedure as shown if FIGS. 23 a and 23 b.
  • FIGS. 23 a and 23 b show the character shift information determination procedure in detail.
  • Referring to FIGS. 23 a and 23 b, the controller 110 verifies an initial appearance position of the character (S601) and determines a movement type of the character (S603). The movement type includes a vertical movement, a horizontal movement, and a diagonal movement. The movement type is randomly selected among the three movement types at step S603. It is assumed that the character moves by 5 pixels.
  • If the movement type is selected, the controller determines whether the movement type is the vertical movement (S605). If the movement type is the vertical movement, the controller 110 determines whether the character's appearance position locates at an upper half part of the screen of the display unit 143 (S607). If the character's appearance position locates at an upper half part of the screen of the display unit 143, the controller 110 generates a character position down shift information ((A, B)=(0, −5)) (S609). If the character's appearance position locates at a lower half part of the screen of the display unit 143, the controller 110 generates a character position up shift information ((A, B)=(0, 5)) (S611).
  • If it is determined, at step S605, that the movement type is not the vertical movement, the controller 110 determines whether the movement type is the horizontal movement (S613). If the movement type is the horizontal movement, the controller determines whether the character's appearance position is located at a right half part of the screen of the display unit 143 (S615). If the character's appearance position locates the right half part of the screen of the display unit 143, the controller generates a character position left shift information ((A, B)=(−5, 0)) (S617). If the character's appearance position is located at a left half part of the screen of the display unit 143, the controller 110 generates character position right shift information ((A, B)=(5, 0)).
  • If it is determined, at step S615, that the movement type is not the horizontal movement, the controller 110 regards the movement type as the diagonal movement.
  • The diagonal movement includes a top-right to bottom-left movement, a top-left to bottom-right movement, a bottom-right to top-left movement, and a bottom-left to top-right movement.
  • If the movement type is the diagonal movement, the controller determines whether the character's appearance position is located at the right half part (X>LCD_X_MAX/2) of the screen (S621). If the character's appearance position is located at the right half part of the screen, the controller determines whether the character's appearance position is located at the upper half part (Y>LCD_Y_MAX/2)) of the screen (S623). If the character's appearance position is located at the upper half part, the controller 110 generates character position left-down shift information ((A,B)=(−5,−5)). If the character's appearance position is not located at the upper half part, the controller 110 generates character position left-up shift information ((A,B)=(−5,5)).
  • If it is determined, at step S621, that the character's appearance position is not located at the right half part of the screen, the controller determines whether the character's appearance position is located at the upper half part (Y>LCD_Y_MAX/2)) of the screen (S629). If the character's appearance position is located at the upper half part, the controller 110 generates character position right-down shift information ((A,B)=(5,−5)) (S631). If the character's appearance position is not located at the upper half part, the controller 110 generates character position right-up shift information ((A,B)=(5,5)) (S633).
  • At step S587 of FIG. 22 a, the controller 110 verifies the appearance position and movement type of the character and calculates the position shift information on the basis of the appearance position and the movement type of the character. The movement direction of the character is changed according to the character's initial appearance position and the movement type of the character is randomly selected among the three movement types (the vertical, horizontal, and diagonal movement). The movement direction is determined such that the character secures the longest length to move at an axis of the center of the screen.
  • If the position shift information is obtained, the controller 10 controls such that the character moves in the movement direction and according to movement type (S589 to S599). The appearance position shift is performed in a predetermined interval. It is assumed that the character moves 20 pixels per second. The character's appearance position shift is performed by 5 pixels every 0.25 seconds. The character is moving and animated at the same time. That is, the character moves in a direction determined at every predetermined times by a unit of predetermined number of pixels. The character can be animated while moving or not moving. The controller 110 determines whether the position shift time arrived (S589). If the position shift time has not arrived, the controller 110 animates the character (S597).
  • If the position shift time has arrived, the controller 110 calculates a next position of the character (S591). The next position is obtained according to the procedure shown in FIGS. 23 a and 23 b. After obtaining the next position, the controller 110 determines whether the next position of the character is in a display range of the screen of the display unit 143 (S593). If the next position of the character is in the display range of the screen, the controller 110 controls to move the character to the next position and animate the character. The controller 110 moves the character on the screen of the display unit 143 in the predetermined movement direction at the predetermined movement times by unit of predetermined number of pixels as shown in FIGS. 4 f and 4 g.
  • If the character deviates from the display range of the screen of the display unit 143 while repeating the above operation, the controller 110 recognizes the deviation of the character and returns to the step S571 to show the idle mode screen as shown in FIG. 4 h. Consequently, the presentations of FIGS. 4 b to 4 g are repeated.
  • The above operation is repeatedly performed while presenting the character house. If the presentation of the character ends (S599), the controller 110 terminates the behavior of the character.
  • The movement of the character is performed after the text expression of the character disappears as shown in FIGS. 4 f and 4 g. The movement direction of the character is randomly selected among the vertical, horizontal, and diagonal direction. That is the movement direction is determined such that the character secures the longest length to move at an axis of the center of the screen and the movement speed is determined at 20 pixels per second. If the character moves across the border of the screen of the display unit 143 to disappear, the character house is presented with the character. After about 2 seconds have passed, the operation depicted in FIG. 4 b is performed.
  • As described above, the presentation of the character behavior in FIGS. 22 a and 22 b is performed by repeating the operations of FIGS. 4 a to 4 g. The character can be presented in 6 steps. The first is an invisible character step as shown in FIGS. 4 a and 4 h, in which the character is not presented yet such that the character house is displayed with the items and background. The second is a character presentation step, in which the character appears at a random position in the character house. When the initial appearance position of the character is close to the border of the screen of the display unit 143 such that the whole image of the character cannot be shown, the character is not presented. The character can appear in a fade-in effect by decreasing the transparency of the character image. The transparency decreased from 100% to 0% (alpha blending image). The third is a sound effect output step as shown in FIG. 4 c, in which a unique sound effect is output according to the state of the character after the character is completely presented. The sound effect is output once. The fourth is an animation step as shown in FIG. 4 d, in which the character is animated after the sound effect is output. The character animation is repeated until the character moves to disappear out of the screen of the display unit 143. The fifth is a text presentation step as shown in FIG. 4 e, in which the state and emotion of the character are presented in text. The text can be presented in the form of a text balloon as shown in FIG. 4 e. The text balloon is presented with a semitransparent balloon with a 30% transparency in consideration of the time taken to read the text. The text balloon disappears after about 3 seconds. The sixth step is a character movement step as shown in FIGS. 4 f to 4 g, in which a movement direction of the character is randomly determined. After the movement direction and distance is determined, the character moves according to the movement direction and distance. The character moves to disappear out of the screen of the display unit 143.
  • Since the character appears at a random position, the character may appear at different position at every appearance time. The movement direction is randomly determined, the character may move in different directions at every movement time. The sound effects and the texts can be differently prepared for expressing various states and emotions of the character. Since the appearance position, movement direction, sound effect, and text are presented differently at every presentation timings, various character behaviors can be provided to the user.
  • At the initial character presentation time, the operations illustrated in FIGS. 4 a to 4 g are performed and then the operations of FIGS. 4 c and 4 e can be skipped after the character moves to disappear out of the screen of the display unit 143. According to an exemplary implementation, the sound effect and the text balloon appear once and the character appears at a random position and moves in a random direction.
  • Referring to FIG. 22 b, the controller 110 verifies information of the current character (S651) and determines a character appearance position (S653). The character appearance position can be a preset position on the screen of the display unit 143. In this embodiment, the character appearance position is preset at a bottom right portion of the screen relative to the center of the screen of the display unit 143. After determining the character appearance position, the controller 110 presents the character with a 100% transparency (S655). Since the transparency is 100%, the character is invisible. Accordingly, the character house is displayed only with the background image and item images as shown in FIG. 4 i.
  • According to an exemplary implementation, the controller 110 decreases the transparency of the character at a predetermined rate such that the character appears gradually. That is, the controller 110 repeats the steps S657 and S659 for gradually presenting the character. If it is assumed that the transparency of the character image decreases 25% every 0.5 seconds, the character is completely presented in 2 seconds (alpha blending effect). The transparency of the character image decreases in an order of 100%, 75%, 50%, 25%, and 0% such that the character house is displayed as shown in FIG. 4 j. After the character has completely appeared, the controller 110 outputs a sound effect corresponding to the state of the character as shown in FIG. 4 c (S661) and then presents a text associated with the state and emotion of the character in the form of a text balloon during a predetermined text presentation time as shown in FIG. 4 l. The text presentation time can be set to about 1.5 seconds in consideration of the time taken for reading the text.
  • After the sound effect is output and after the text has been presented, the controller 110 presents a character animation at steps S665 to S669. The character animation is repeated until the character disappears from the character house. As described above, the character presentation is performed in a specific order according to the character appearance, output of sound effects, presenting text balloon, and character animation.
  • As described above, the screen image presentation method of an exemplary embodiment of the present invention presents the character house (screen) in a specific operation (idle mode) of the mobile phone. In the idle mode, the controller of the mobile phone presents a character house by arranging a background and items in the character house and presents a character. The character expresses a state and emotion in the forms of a sound effect and text. The character moves in random direction and shows an animation on the screen of the display unit 143. The background image is used as a background of the house and can be replaced with another image to graphically illustrate the progress of the time using character house. The items presented with the character house includes movement items, decoration items, and function items. The function items are used for improving user convenience by showing system information of the mobile phone. The decoration items are used for improving the aesthetic effect of the screen of the mobile phone. The motion items are used to make dynamic the screen by moving around the screen house. Also, the character house can be implemented with a special expression function such that the character gives special information at predetermined time zones or greets to the user. Also, the character house can be implemented such that the character gives an item or character money as a gift at unexpected timing while the character greets, resulting in user's enjoyment.
  • In this exemplary embodiment of the present invention, the character house is structured by presenting the elements in an order of the background, items, and character. However, the presentation order can be changed. In this exemplary embodiment of the present invention, the character house is configured such that the greeting and event presentation are performed when the item is presented, however, the greeting and event presentation can be performed when the background or character is presented.
  • FIG. 24 is a flowchart illustrating a screen house presentation procedure of a screen image presentation method for a mobile phone according to another exemplary embodiment of the present invention.
  • In association with the character house presentation, the mobile phone can be in a character house configuration mode, character house editing mode, or character house presentation mode.
  • The character house configuration mode and character house editing mode can be accessed by a corresponding menu selection or a specific key (including a soft key) manipulation, and the character house presentation mode activates for a specific operation mode of the mobile terminal when the character house presentation function is enabled.
  • Referring to FIG. 24, if a character house configuration mode is executed by a mode execution key or menu selection, the controller 110 detects the execution of the character house configuration mode (S719) and then displays a character house configuration menu screen for allowing a user to configure the character house presentation options (S721). The character options include a character house presentation on/off option and character quotients presentation on/off option. If the character house presentation on/off option is set to off, the controller 110 disables the character house presentation function. If the character quotients presentation on/off option is set to off, the controller 110 disables the presentation of the character quotients shown in FIGS. 7 b and 8. The character house option may include a greeting function on/off option. If the greeting function on/off option is set to off, the controller 110 disables the greeting behavior presentation shown in FIG. 12 and greeting event presentation shown in FIG. 13.
  • If the character house editing mode is executed by a mode executing key or menu selection, the controller 110 detects the executing of the character house editing mode (S715) and then displays a character house editing menu screen for allowing the user to edit the character house (S717). In the character house editing mode, it is possible to change, add, and delete the background, items, and character.
  • FIG. 25 is a state diagram illustrating a character house design procedure of a screen image presentation method according to an exemplary embodiment of the present invention.
  • Referring to FIG. 25, the controller 110 displays a character house configuration screen in the character house configuration mode 810, in which it is allowed to delete a character house or to set a default character house. If a “new character house” option is selected or a “character house editing” option is selected on the character house configuration screen, the controller 110 displays a character house editing screen in the character house editing mode 820. From the character house editing mode 820, it is possible to enter modes 830 to 870 for editing a background and items.
  • FIG. 26 is a state diagram illustrating transitions between the character house configuration mode and the character house editing mode of FIG. 25. FIG. 27 is a screen image illustrating a character house configuration screen in the character house configuration mode of FIG. 25.
  • Referring to FIGS. 26 and 27, the character house configuration screen displays character houses that can be used for a specific mode screen image. The mobile phone includes a plurality of character houses. These character houses are displayed on the character house configuration screen in the form of tiled windows as shown in FIG. 27. In this exemplary embodiment of the present invention, the character house configuration screen is assumed to display four character houses at once. If there are three character houses stored in the mobile phone, one of the four tiled window is shown empty. In relation with the empty character house window, it is impossible to set a default character house, delete a character house, and edit a character house. Also, a new character house can be put into the empty window. If an OK key is input while the empty window is selected, a character creation function is enabled. If the OK key is input while one of the tiled windows having the character houses is selected, the character house is set for a screen image for a specific operation mode of the mobile phone.
  • In order to select one of the character houses on the character house configuration screen, the cursor navigates on the tiled windows in accordance with a direction key (navigation key) manipulation. When the cursor is positioned on one of the tiled windows, the controller 110 displays a menu having options shown in table 4.
    TABLE 4
    Empty window 2. new character house
    Non-empty window 1. set default
    2. edit character house
    3. delete character house
  • FIG. 28 is a state diagram illustrating transitions between the character house configuration mode and the character house editing mode of FIG. 25, and FIG. 29 shows screen images illustrating a procedure for setting a “character house presentation on/off” option on the character house configuration screen of FIG. 25.
  • Referring to FIGS. 28 and 29. the “character house presentation on/off” option can be set to “on” or “off.” The controller 110 displays the character houses on the character house configuration screen in the form of tiled windows as shown in a first screen image of FIG. 29. If one of the windows is selected while the “character house presentation on/off” option is set to off, the controller 110 highlights the selected window and displays an announcement message asking whether to use the character house presentation function as shown in a second screen image of FIG. 29. If a user command is input for using the character house presentation function, the controller 110 enables the character house presentation function and displays the character house of the selected window as shown in a third screen image of FIG. 29.
  • As described above, if the character house configuration screen is loaded while the character house presentation function is disabled, the controller 110 displays an announcement message asking whether to use the character house presentation function and then enables the character house presentation function according to the user command for using the character house presentation function.
  • FIG. 30 is a state diagram illustrating transitions among the character house editing mode and specific item editing modes of FIG. 25, and FIG. 31 shows a screen image illustrating an example character house editing screen in the character house editing mode of FIG. 25.
  • Referring to FIGS. 30 and 31, the character house editing screen is displayed when a “new character house” option or “edit character house” option is selected on the character house configuration screen. The character house editing screen provides a menu with options for editing the background and item (1. Background, 2. New item). When the “new character house” option is selected, a selected background is basically presented without items. The character house editing screen displays a key input guide as shown in FIG. 31. If a menu option is selected, the menu with options for editing the background and items is displayed.
  • FIG. 32 is a state diagram illustrating transitions between the character house editing mode and the background selection mode of FIG. 25. FIG. 33 shows screen images illustrating a procedure for selecting a background on the background selection mode of FIG. 32.
  • Referring to FIGS. 32 and 33, if a “background” option is selected on the character house editing screen, the controller 110 controls to present the background selection screen in a background selection mode 830 as shown in a first screen image of FIG. 33. The background selection screen enables navigating background images using direction keys (navigation keys). If a selection key (OK key) is input while one of the background images is presented, the controller 110 returns to display the character house editing mode 820 and displays the selected background image as the background on the screen of the display unit 143 as shown in a second screen image of FIG. 33.
  • As described above, if the “background” option is selected on a character house editing screen, the controller 110 presents available background images stored in the mobile phone with titles of the background images. The background images are individually shown according to the navigation key manipulation. If a selection key is input while a background image is displayed, the background image is stored as the background and the character house editing screens are redisplayed.
  • FIG. 34 is a state diagram illustrating transitions between the character house editing mode and item editing modes of FIG. 25. FIGS. 35 a to 35 d illustrate screen images illustrating a procedure for adding an item to the character house in the character house editing mode of FIG. 34.
  • Referring to FIGS. 34 and 35 a to 35 b, if a “new item” option is selected from option menu on the character house editing screen, as shown in FIG. 35 a, in the character house editing mode 820, the controller 110 displays a new item selection screen in a new item selection mode 860 as shown in FIG. 35 b. The new item selection screen enables navigating item images using direction keys (navigation keys). Each item image is presented with a title of the item image.
  • If a selection key (OK key) is input during the presentation of one of the item images, the controller 110 displays an item positioning screen, as shown in FIG. 35 c, in an item positioning mode 870. The item positioning screen is partitioned to define a plurality of sections on the screen of the display unit 143 and each section is assigned a unique section number. The item positioning screen facilitates the movement of an item from one place to another with reference to the section number. If a section number is input while the item positioning screen is displayed, the controller 110 displays a selected item editing screen, in which the selected item is positioned at the section having the input section number as shown in FIG. 35 d, in a selected item editing mode 850. The selected item editing screen facilitates fine adjustment of the selection item position by manipulating the navigation keys.
  • The selected item editing screen also provides an item-fixing key (for example, # key) for fixing the selected item at the adjusted position and a fast move key (for example * key) for quickly moving the item to another section in the item positioning mode 870. If an additional item is selected from the option menu of the selected item editing mode 850, the controller 110 displays the new item selection screen of the new item selection mode 860 so as to allow selecting another item. If the fast move key is input while the selected item editing screen is presented, the controller 110 performs displays the item position screen of the item position mode 870 so as to move the item to another section.
  • As described above, the selected item editing mode 850 facilitates fine adjustment of the position of the selected item on the selected item editing screen. The items can be overlapped each other. The overlapped presentation of the items can be performed with a “send to front,” “send to back,” and “message adjustment options displayed when an option key is input. If the “send to front” (a “volume up” key) or the “send to back” (a “volume down” key), the controller 110 places the item on or beneath the item in the selected section.
  • If a fast move key (* key) is selected while the mobile phone is in the selected item editing mode 850, the mobile phone enters the item positioning mode 870. If the item-fixing key is selected, the mobile phone enters an edited item shortcut selection mode 840. If an “add item” option is selected, the mobile phone enters the new item selection mode 860. If a “delete item” option is selected, the mobile phone enters the character house editing mode 820. If a “background” option is selected, the mobile phone enters the background selection mode 830. The transition routes from the selected item editing mode are summarized as shown in table 5.
    TABLE 5
    Functions in screen Fine position adjustment function (navigation key)
    having a selected Item overlapping display function (option >>
    item to be edited send to front, send to back, background)-
    background selection screen
  • FIG. 36 is a state diagram illustrating transitions between the character house editing mode and the selected item editing mode via the edited item shortcut selection mode of FIG. 25, and FIGS. 37 a to 37 c show screen images illustrating a procedure for assigning shortcut key in the edited item shortcut selection mode of FIG. 36.
  • Referring to FIGS. 36 and 37, an item to be edited is selected on the edited item shortcut selection screen in the edited item shortcut selection mode 840. The edited item shortcut selection mode 840 is transited by from the character house editing mode 820 represented by the character house editing screen as shown in FIG. 37 a and the selected item editing mode 850 represented by the selected item editing screen as shown in FIG. 37 b. If the character house editing screen is open, the controller 110 displays the items included in the character house as shown in FIG. 37 b. A character house can be provided with maximal 10 items that are numbered from 0 to 9. Accordingly, if an item fixing key (#) is input while the character house editing screen is displayed, the controller 110 displays the edited item shortcut selection screen to assign the section numbers to the items. If a section number of an item is in the edited item shortcut selection screen, the controller 110 displays the selected item editing screen to display the selected item in a bracket as shown in FIG. 37 c. The selected item in the bracket can be finely adjusted in position. If the fast move key (*) is selected to quickly move the selected item at other section, the controller 110 displays the item positioning screen defining the sections.
  • As described above, the screen image presentation method for a mobile phone according to an exemplary embodiment of the present invention facilitates the design of a character house that is used as a screen image in a specific operation mode of the mobile phone. The character house can be edited by adding, deleting, and changing positions of the background and items.
  • The screen image presentation method provides the mobile phone with the character house editing mode 820 in which the character houses stored in the mobile phone can be selected, deleted, and edited. The mobile phone can switch between the character house-related editing modes for creating a new character house.
  • The background selection mode 830 is represented by the background selection screen which is accessed from the selected item editing screen of the selected item editing mode 850 or by selecting the “background” option in the menu provided in the selected item editing mode 850. The controller 110 presents the background image together with the title of the background image as shown in the first screen image of FIG. 33. The background image slides according to the direction key (left and right direction keys) manipulation to make the selection of one of the background images as the background of the character house possible.
  • The new item selection mode 860 facilitates the addition of a new item to the character house and transitions to the item positioning mode 870 and the selected item editing mode 850. If the “new item” option is selected in the menu provided in the character house editing mode 820, the controller 110 displays the new item selection screen, as shown in FIG. 35 b, representing the new item selection mode 860. The controller 110 presents the item images together with the title of the items. The new item selection screen facilitates the navigation of item images using direction keys (left and right direction keys). If one of the items is selected from the new item selection screen, the controller 110 displays the item positioning screen, as shown in FIG. 35 c, representing the item positioning mode 870. The item positioning screen 870 is partitioned so as to define several sections as shown in FIG. 35 c. In this exemplary embodiment of the present invention, the screen is divided into 9 sections and assigned unique section numbers, respectively. If a section number is selected while the item positioning screen, the controller 110 displays the selected item editing screen as shown in FIG. 35 d. The selected item editing screen facilitates fine adjustment of the position of the selected item. If the addition of another item is required, the controller 110 displays the new item selection screen again, i.e. enters the new item selection mode 860.
  • As described above, the controller 110 presents the items stored in the mobile phone if there is a command input for adding an item to the character house. The screen is divided into several sections for quickly positioning an item on the screen and the item is finely adjusted in a section.
  • The controller 110 operates the edited item shortcut mode 840 and the selected item editing mode 850 to facilitate a change in the position of the item in the character house. That is, the controller 110 assigns the item numbers to the items presented with the character house on the edited item shortcut screen and highlights the item selected by the item number as a target item. The target item is edited in the selected item editing mode 850 and the item positioning mode 870.
  • As described above, the controller 110 facilitates the editing of the character house at the steps 719 and 721 of FIG. 24. In these steps, the controller 110 also determines whether to present the character quotients and the state and emotion of the character. Also, the controller 110 facilitates the setting of whether to activate the greeting behavior and random greeting event of the character.
  • The controller 110 also allows editing the background and items and adjusting the position of the items at steps 715 and 717 of FIG. 24. As described above, the character house includes the character, background, and items. The character house can be created in the mobile phone and store at least one character house. The controller 110 of the mobile phone facilitates the selection of one of the character houses in order for a selected character house to be used as a screen image in a specific operation mode of the mobile phone. The background and items constituting the character house can be edited. The background of the character house can be deleted, added, or modified in the character house editing mode under the control of the controller 110. Also, the items of the character house can be deleted or added, and repositioned on the screen in the character house editing mode.
  • If the character house is created with the character, background, and items, and the character house presentation function is activated, the controller 110 of the mobile phone performs the operations for presenting the character house at steps S711 and S713 of FIG. 24. The character house presentation is performed in the idle mode of the mobile phone at the same or a different time with a start of the idle mode. That is, the character house presentation can be set to start at the time when the display unit 143 is turned on by opening the clam shell phone or sliding up open the slide phone, or after completing an event of the mobile phone. The character house presentation also can be set to start when a special event occurs. For example, the character house can be presented when verifying an incoming message, writing an outgoing message, or verifying a phone book.
  • If the message or information is displayed with the character house, the controller 110 controls such that the character house is presented in a transparency higher than that of the message or information, using the alpha blending method. For, example, the character house is presented in 50% transparency while the message or information in 0% transparency. Also, the character house can be implemented with a display window such the message or information is presented in the display window of the character house. The character house is preferably presented during a preset time to reduce power consumption of the mobile phone.
  • If the mobile phone is in the character house presentation mode (S711), the controller 110 controls to present the character house (S713).
  • The character house is presented while the controller 110 performs the steps S313 to 315 of FIG. 14. That is, the controller 110 verifies the character house configuration information (S313) and determines whether the character house function is turned on (S315). If the character house function is turned on, the controller 110 verifies the character house information (S319) and then presents the character house in accordance with the character house information. As described above, the character house includes a background, items, and character. The controller 110 presents the background of the character house according to the procedure of FIG. 15. If the background is a time-varying background, the controller 110 verifies the current time and presents a background image corresponding to the current time. The controller 110 presents the items of the character house according to the procedure of FIG. 16. The controller 110 determines whether the character quotients presentation function is turned on. If the character quotients presentation function is turned on, the controller presents the character quotients. The controller 110 analyzes the types of items and presents the items according to the procedures corresponding to the types of the items. If the item is a motion item, the motion item is presented according to the procedure of FIG. 17. If the item is a function item, the function item is presented according to the procedure of FIG. 18. The presentation of the character is performed according to the procedure of FIG. 19. If the greeting behavior function is set to “on,” the controller 110 verifies the current time and controls the character to show the greeting behavior if the current time is in a greeting time zone of a day. If a greeting event occurs while the character shows the greeting behavior, the controller 110 controls to preset the greeting event according to the procedure of FIG. 21. The behavior of the character house is performed according to the procedure of FIG. 22 and a character movement is performed according to the procedure of FIG. 23.
  • The controller 110 facilitates the setting of the character house functions, editing the character house, and presenting the character house according to the preset functions, according to the procedure of FIG. 24.
  • As described above, the character house of the present invention is a virtual space in which an artificial intelligent character resides, talks, and moves such that a user recognize a state of a mobile by means of the character house. The character house facilitates learning functions of the mobile phone and attracts the user's interest to the mobile phone. The character house of an exemplary embodiment of the present invention is presented with a character, background image, items, and animation. The character can present a sound effect and text expression reflecting its condition and emotion according to a character scenario, moves around the character house, and shows an animation. The character can express the state and emotion of the mobile phone in various manners so as to make the character house dynamic. The background is used to lively express the character house by changing the background image according to a progress of the time, whereby the user can guess the current time from the character house. The items includes decoration items, motion items, and function items. The function items show system information on the mobile phone to improve user convenience. The decoration items increase the aesthetic sense of the character house. The motion items move around in the character so as to make the character house dynamic.
  • The character house can be implemented with a special expression method in which the character greets the user of the mobile phone with an appropriate sound, text, and behavior at predetermined time zones of a day, to make the mobile phone user-friendly. Also, the character house can be implemented such that the character randomly gives a gift to the user, resulting in enjoyment of the user.
  • While the present invention has bee shown and described with reference to certain exemplary embodiments thereof, it will be understood by those skilled in the art that various changes in form and details may be made therein without departing from the spirit and scope of the invention as defined by the appended claims and their equivalents.

Claims (51)

1. A screen image presentation apparatus for a mobile phone, comprising:
a memory for storing a character house comprising a character for presenting a condition of the mobile phone;
a controller for controlling the presentation of the character house at a presentation time; and
a display for displaying the character house under the control of the controller.
2. The screen image presentation apparatus of claim 1, wherein the character house comprises:
items arranged in the character house; and
a background arranged behind the character house.
3. The screen image presentation apparatus of claim 2, wherein the items comprise function items that represent conditions of specific functions of the mobile phone and illustrate respective menus when selected.
4. The screen image presentation apparatus of claim 3, wherein the function items comprise:
a message item for presenting a number of unread messages;
an alarm clock item for presenting scheduled alarms; and
a memory monitor item for presenting memory utilization.
5. The screen image presentation apparatus of claim 3, wherein the items comprise:
motion items for moving along movement paths in the character house; and
decoration items for decorating the character house at fixed positions in the character house.
6. The screen image presentation apparatus of claim 3, wherein the background comprises a plurality of background images changing in accordance with the progress of time.
7. The screen image presentation apparatus of claim 2, wherein the controller controls presentation of the character house when the mobile phone operates in an idle mode.
8. The screen image presentation apparatus of claim 7, wherein the controller controls presentation of the character house in a time sequence of the background, items and character.
9. The screen image presentation apparatus of claim 8, wherein the controller verifies the current time and presents a background image corresponding to the current time.
10. The screen image presentation apparatus of claim 8, wherein the controller verifies the types of the items and presents the items differently according to the type of the item, the types of the items comprising decoration items that are fixed at positions of the character house, motion items that are moving along movement paths in the character house, and the function items that illustrate conditions of the mobile phone.
11. The screen image presentation apparatus of claim 10, wherein the controller analyzes a condition of a function represented by the function item and presents the function item together with an analysis result in the form of a popup window.
12. The screen image presentation apparatus of claim 11, wherein the function items comprise a message item presented together with the number of unread messages in the form of a popup window.
13. The screen image presentation apparatus of claim 11, wherein the function item comprises an alarm item presented together with scheduled alarms in the form of the popup window.
14. The screen image presentation apparatus of claim 11, wherein the motion item moves along the path in units of time.
15. The screen image presentation apparatus of claim 10, wherein the controller determines a position of the character in the character house and animates the character.
16. The screen image presentation apparatus of claim 15, wherein the controller starts presenting the character in a transparency, outputs sound effects and texts according to progression of the transparency, and animates the character.
17. The screen image presentation apparatus of claim 16, wherein the controller verifies the current time and presents a greeting sound and text if the current time is in at least one of greeting times.
18. A screen image presentation method for a mobile phone, comprising:
verifying configuration information on a character house comprising a character representing a condition that is determined by usage pattern of the mobile phone; and
presenting the character house in a specific operation state of the mobile phone on the basis of the configuration information.
19. The screen image presentation method of claim 18, wherein the character house comprises:
items arranged in the character house; and
a background arranged behind the character house.
20. The screen image presentation method of claim 19, wherein the specific operation state comprises an idle state of the mobile phone.
21. The screen image presentation method of claim 20, wherein the character house is presented in a time sequence of the background, items, and character.
22. The screen image presentation method of claim 21, wherein the items comprise function items that represent conditions of specific functions of the mobile phone and illustrate respective menus when selected.
23. The screen image presentation method of claim 22, wherein the items further comprise:
motion items that move along movement paths in the character house; and
decoration items for decorating the character house at fixed positions in the character house.
24. The screen image presentation method of claim 23, wherein the background comprises a plurality of background images changing in accordance with the progress of time.
25. The screen image presentation method of claim 23, wherein presenting the character house comprises:
determining whether an item comprises a function item:
determining whether the function item is set to operating if an item comprises a function item; and
presenting the function item together with information collected by a function represented by the function item in the form of a popup window if the function item is set to operating.
26. The screen image presentation method of claim 25, wherein the function items comprise a message item which is presented together with the number of unread messages in the form of the popup window.
27. The screen image presentation method of claim 25, wherein the function items comprise an alarm item which is presented together with scheduled alarms in the form of a popup window.
28. The screen image presentation method of claim 23, wherein each motion item moves along the path in units of movement time.
29. The screen image presentation method of claim 23, wherein presenting the character house comprises:
determining a position of the character in the character house; and
animating the character at the position.
30. The screen image presentation method of claim 23, wherein presenting the character house comprises:
starting presentation of the character at an appearance position in a transparency;
outputting sound effects and texts according to progressing the transparency; and
animating the character at the appearance position.
31. The screen image presentation method of claim 29, wherein presenting the character house further comprises:
determining whether a current time is within at lest one of a preset greeting time zone of a day; and
presenting a greeting sound and text if the current time is within one of the preset greeting time zones.
32. The screen image presentation method of claim 21, wherein presenting the character house comprises:
verifying the current time; and
presenting a background image corresponding to at least one of a preset time zone to which the current time belongs.
33. The screen image presentation method of claim 21, wherein presenting the character house comprises:
determining whether an item comprises a function item;
presenting a function item image corresponding to the function item at a position in the character house together with information collected by a function represented by the function item in the form of a popup window, if an item comprises a function item;
determining whether the function item comprises a motion item if the item is not a function item;
presenting a motion item image corresponding to the motion item and moving the motion item along a path in units of movement time if the item comprises a motion item; and
determining that the item comprises a decoration item and presenting the decoration item at a position in the character house if the item does not comprise a motion item.
34. The screen image presentation method of claim 33, wherein presenting the character house further comprises:
determining whether the function item comprises a mailbox item if the item comprises a function item;
verifying a number of unread messages and presenting the number of the unread messages together with the mailbox item if the function item comprises a mailbox item;
determining whether the function item comprises an alarm item if the function item does not comprise a mailbox item;
verifying scheduled alarms and presenting the alarm scheduled at the closest scheduled alarm time together with the alarm item if the function item comprises an alarm item;
determining that the function item comprises a memory item if the function item does not comprise an alarm item; and
verifying a memory utilization and presenting the memory utilization together with the memory item.
35. The screen image presentation method of claim 33, wherein presenting the character house further comprises:
determining whether a character information presentation function is set to operating;
verifying character quotients if a character information presentation function is set to operating;
determining an emotion representing the character quotients;
presenting an emoticon and text corresponding to the emotion; and
displaying the character quotients in a character information presentation region.
36. The screen image presentation method of claim 21, wherein presenting the character house comprises:
presenting the character to gradually appear by reducing a transparency of the character in preset steps after the background and the items are presented;
presenting a sound and text corresponding to a condition of the mobile phone; and
moving the character in a reference direction.
37. The screen image presentation method of claim 36, wherein presenting the character comprises:
verifying the current time;
determining whether the character has been presented in a preset time zone to which the current time belongs; and
presenting a greeting behavior of the character if the character is presented in a time zone to which the current time belongs.
38. The screen image presentation method of claim 37, wherein presenting a greeting behavior comprises:
animating the character;
presenting a text, in the form of a text balloon;
notifying that the greeting behavior corresponding to the preset time zone is performed;
determining whether the current time comprises a greeting event time; and
illustrating the greeting event if the current time comprises a greeting event time.
39. The screen image presentation method of claim 38, wherein showing the greeting event comprises:
determining at least one of an item and character money to present to the user as a gift;
presenting the at least determined item and character money; and
presenting a greeting event text in the form of a text balloon together with an animation of the character.
40. A screen image presentation method for a mobile phone, comprising:
registering configuration information of a character house in a program mode;
editing the character house comprising a background and items in an editing mode;
verifying the configuration information of the character house; and
presenting the character house with a character, items, and background on the basis of the configuration information of the character house.
41. The screen image presentation method of claim 40, wherein registering configuration information comprises setting a character house presentation on/off option and a character information presentation on/off function.
42. The screen image presentation method of claim 41, wherein editing the character house comprises:
illustrating a background image together with a title while editing the background of the character house;
registering the background image as the background of the character house;
adding an item in an item editing mode;
presenting a plurality of sections defined by dividing the character house;
assigning section numbers to the sections;
positioning the item in any of the sections by selecting the corresponding section numbers; and
adjusting a position of the item finely in the section.
43. The screen image presentation method of claim 42, wherein adding an item comprises:
displaying an item image and title of the item image;
showing the sections with section numbers on the character house;
presenting, if a section number is selected, the item in the section to which the section number is assigned;
adjusting the position of the item finely in the section; and
registering a type and position of the item in the character house.
44. The screen image presentation method of claim 42, wherein adjusting a position of the item comprises:
defining sections by partitioning the character house if an item to be edited is selected and assigning section numbers to the sections;
positioning, if an item number is selected, the item in the section corresponding to the section number;
adjusting the position of the item finely in the section; and
registering a type and position of the item in the character house.
45. The screen image presentation method of claim 40, wherein presenting the character house comprises:
determining whether the character house presentation on/off option is set for operating;
extracting identification and position information of the background and items of the character house from the configuration information of the character house if the character house presentation on/off option is set for operating; and
presenting the background and items on the basis of the identification and position information and then presenting the character.
46. A screen image presentation apparatus for a mobile phone, comprising:
a memory for storing an artificial intelligence-based character house in which an artificial-intelligence-based character resides for presenting a condition of the mobile phone;
wherein the character house comprises items arranged in the character house and a background arranged behind the character house;
a controller for controlling the presentation of the character house at a presentation time; and
a display for displaying the character house under the control of the controller.
47. The screen image presentation apparatus of claim 46, wherein the items comprise function items that represent conditions of specific functions of the mobile phone and illustrate respective menus when selected.
48. The screen image presentation apparatus of claim 47, wherein the function items comprise:
a message item for presenting a number of unread messages;
an alarm clock item for presenting scheduled alarms; and
a memory monitor item for presenting memory utilization
49. The screen image presentation apparatus of claim 47, wherein the items comprise:
motion items for moving along movement paths in the character house; and
decoration items for decorating the character house at fixed positions in the character house.
50. The screen image presentation apparatus of claim 47, wherein the background comprises a plurality of background images changing in accordance with the progress of time.
51. The screen image presentation apparatus of claim 8, wherein the controller verifies the types of the items and controls to present the items differently according to at leat one of the type of the item, the types of the items comprising decoration items that are fixed at positions of the character house, motion items that are moving along movement paths in the character house, and the function items that illustrate conditions of the mobile phone.
US11/633,624 2005-12-06 2006-12-05 Screen image presentation apparatus and method for mobile phone Abandoned US20070171192A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
KR20050118428 2005-12-06
KR2005-0118428 2005-12-06

Publications (1)

Publication Number Publication Date
US20070171192A1 true US20070171192A1 (en) 2007-07-26

Family

ID=37629899

Family Applications (2)

Application Number Title Priority Date Filing Date
US11/633,641 Expired - Fee Related US8132100B2 (en) 2005-12-06 2006-12-05 Screen image presentation apparatus and method for mobile phone
US11/633,624 Abandoned US20070171192A1 (en) 2005-12-06 2006-12-05 Screen image presentation apparatus and method for mobile phone

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US11/633,641 Expired - Fee Related US8132100B2 (en) 2005-12-06 2006-12-05 Screen image presentation apparatus and method for mobile phone

Country Status (4)

Country Link
US (2) US8132100B2 (en)
EP (2) EP1796355A3 (en)
KR (3) KR100785067B1 (en)
CN (2) CN1980441B (en)

Cited By (72)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080119261A1 (en) * 2006-10-23 2008-05-22 Jorge Heymann Slot machine bonus game providing awards for manual dexterity
US20090124352A1 (en) * 2007-11-13 2009-05-14 Ignacio Gerson Slot machine game with side pot
US20090199110A1 (en) * 2008-02-05 2009-08-06 Samsung Electronics Co., Ltd. Apparatus and method for transmitting animation-based message
US20100107068A1 (en) * 2008-10-23 2010-04-29 Butcher Larry R User Interface with Parallax Animation
US20100105424A1 (en) * 2008-10-23 2010-04-29 Smuga Michael A Mobile Communications Device User Interface
US20100222112A1 (en) * 2009-03-02 2010-09-02 Seung Sook Han Method for displaying items in mobile terminal and mobile terminal using the same
US20100248787A1 (en) * 2009-03-30 2010-09-30 Smuga Michael A Chromeless User Interface
US20100248688A1 (en) * 2009-03-30 2010-09-30 Teng Stephanie E Notifications
WO2010117642A3 (en) * 2009-03-30 2011-01-13 Microsoft Corporation Unlock screen
US20110283238A1 (en) * 2010-05-12 2011-11-17 George Weising Management of Digital Information via an Interface
US20120026174A1 (en) * 2009-04-27 2012-02-02 Sonoma Data Solution, Llc Method and Apparatus for Character Animation
US20120151381A1 (en) * 2010-12-14 2012-06-14 Microsoft Corporation Defining actions for data streams via icons
US20120165071A1 (en) * 2010-12-28 2012-06-28 Inventec Appliances (Shanghai) Co. Ltd. Mobile device capable of automatically switching its operation modes
CN102622214A (en) * 2011-01-27 2012-08-01 腾讯科技(深圳)有限公司 Method and device for realizing multiple-display mode universal icons
US8269736B2 (en) 2009-05-22 2012-09-18 Microsoft Corporation Drop target gestures
US8385952B2 (en) 2008-10-23 2013-02-26 Microsoft Corporation Mobile communications device user interface
US8411046B2 (en) 2008-10-23 2013-04-02 Microsoft Corporation Column organization of content
US20130088516A1 (en) * 2010-05-17 2013-04-11 Ntt Docomo, Inc. Object displaying apparatus, object displaying system, and object displaying method
US20130262511A1 (en) * 2012-04-02 2013-10-03 Google Inc. Determining 3D Model Information From Stored Images
US8560959B2 (en) 2010-12-23 2013-10-15 Microsoft Corporation Presenting an application change through a tile
US8689123B2 (en) 2010-12-23 2014-04-01 Microsoft Corporation Application reporting in an application-selectable user interface
US8687023B2 (en) 2011-08-02 2014-04-01 Microsoft Corporation Cross-slide gesture to select and rearrange
US8690664B2 (en) 2006-09-25 2014-04-08 Etasse Limited Slot machine game with additional award indicator
US8702493B2 (en) 2007-11-09 2014-04-22 Etasse Limited Slot machine game with award based on another machine
US20140195144A1 (en) * 2013-01-10 2014-07-10 Fujitsu Limited Analysis method, and analysis apparatus
US8830270B2 (en) 2011-09-10 2014-09-09 Microsoft Corporation Progressively indicating new content in an application-selectable user interface
US8836648B2 (en) 2009-05-27 2014-09-16 Microsoft Corporation Touch pull-in gesture
US8893033B2 (en) 2011-05-27 2014-11-18 Microsoft Corporation Application notifications
US8922575B2 (en) 2011-09-09 2014-12-30 Microsoft Corporation Tile cache
US8933952B2 (en) 2011-09-10 2015-01-13 Microsoft Corporation Pre-rendering new content for an application-selectable user interface
US8935631B2 (en) 2011-09-01 2015-01-13 Microsoft Corporation Arranging tiles
US8990733B2 (en) 2010-12-20 2015-03-24 Microsoft Technology Licensing, Llc Application-launching interface for multiple modes
US9052820B2 (en) 2011-05-27 2015-06-09 Microsoft Technology Licensing, Llc Multi-application environment
US9104440B2 (en) 2011-05-27 2015-08-11 Microsoft Technology Licensing, Llc Multi-application environment
US9128605B2 (en) 2012-02-16 2015-09-08 Microsoft Technology Licensing, Llc Thumbnail-image selection of applications
US20150278765A1 (en) * 2014-04-01 2015-10-01 Microsoft Corporation Information collections
US9158445B2 (en) 2011-05-27 2015-10-13 Microsoft Technology Licensing, Llc Managing an immersive interface in a multi-application immersive environment
US20150370448A1 (en) * 2011-11-28 2015-12-24 Kyocera Corporation Device, method, and storage medium storing program
US9223472B2 (en) 2011-12-22 2015-12-29 Microsoft Technology Licensing, Llc Closing applications
US9244802B2 (en) 2011-09-10 2016-01-26 Microsoft Technology Licensing, Llc Resource user interface
US20160080298A1 (en) * 2014-09-12 2016-03-17 Samsung Electronics Co., Ltd. Method for generating emoticon and electronic device supporting the same
US9329774B2 (en) 2011-05-27 2016-05-03 Microsoft Technology Licensing, Llc Switching back to a previously-interacted-with application
US20160125715A1 (en) * 2014-10-30 2016-05-05 Fih (Hong Kong) Limited Electronic device and alarm method of the electronic device
US9383917B2 (en) 2011-03-28 2016-07-05 Microsoft Technology Licensing, Llc Predictive tiling
US9423951B2 (en) 2010-12-31 2016-08-23 Microsoft Technology Licensing, Llc Content-based snap point
US9430130B2 (en) 2010-12-20 2016-08-30 Microsoft Technology Licensing, Llc Customization of an immersive environment
US9450952B2 (en) 2013-05-29 2016-09-20 Microsoft Technology Licensing, Llc Live tiles without application-code execution
US9451822B2 (en) 2014-04-10 2016-09-27 Microsoft Technology Licensing, Llc Collapsible shell cover for computing device
US20160352667A1 (en) * 2015-06-01 2016-12-01 Facebook, Inc. Providing augmented message elements in electronic communication threads
US9520031B2 (en) 2008-07-07 2016-12-13 Etasse Limited Slot machine game with symbol lock-in
US9557909B2 (en) 2011-09-09 2017-01-31 Microsoft Technology Licensing, Llc Semantic zoom linguistic helpers
US20170052946A1 (en) * 2014-06-06 2017-02-23 Siyu Gu Semantic understanding based emoji input method and device
US9658766B2 (en) 2011-05-27 2017-05-23 Microsoft Technology Licensing, Llc Edge gesture
US9665384B2 (en) 2005-08-30 2017-05-30 Microsoft Technology Licensing, Llc Aggregation of computing device settings
US9674335B2 (en) 2014-10-30 2017-06-06 Microsoft Technology Licensing, Llc Multi-configuration input device
US9684284B2 (en) 2015-03-27 2017-06-20 Alibaba Group Holding Limited Setting an alarm clock on a smart device
US9769293B2 (en) 2014-04-10 2017-09-19 Microsoft Technology Licensing, Llc Slider cover for computing device
US9841874B2 (en) 2014-04-04 2017-12-12 Microsoft Technology Licensing, Llc Expandable application representation
US20180024714A1 (en) * 2012-04-04 2018-01-25 Samsung Electronics Co., Ltd. Intelligent event information presentation method and terminal
US10068256B2 (en) 2014-10-08 2018-09-04 Microsoft Technology Licensing, Llc User directed information collections
US10254942B2 (en) 2014-07-31 2019-04-09 Microsoft Technology Licensing, Llc Adaptive sizing and positioning of application windows
US10353566B2 (en) 2011-09-09 2019-07-16 Microsoft Technology Licensing, Llc Semantic zoom animations
US10531852B2 (en) * 2013-12-24 2020-01-14 Samsung Electronics Co., Ltd. Display for converting medical images and method of use
US10592080B2 (en) 2014-07-31 2020-03-17 Microsoft Technology Licensing, Llc Assisted presentation of application windows
US10642365B2 (en) 2014-09-09 2020-05-05 Microsoft Technology Licensing, Llc Parametric inertia and APIs
US10678412B2 (en) 2014-07-31 2020-06-09 Microsoft Technology Licensing, Llc Dynamic joint dividers for application windows
WO2020128962A1 (en) * 2018-12-19 2020-06-25 RxPrism Health Systems Pvt. Ltd A system and a method for creating and sharing interactive content rapidly anywhere and anytime
JP2021139904A (en) * 2019-12-24 2021-09-16 株式会社ユピテル Device and program
US11350007B2 (en) * 2018-12-07 2022-05-31 Canon Kabushiki Kaisha Image reading apparatus, image processing method, and storage medium for correcting show-through
US20220245681A1 (en) * 2020-01-10 2022-08-04 Tencent Technology (Shenzhen) Company Limited Item display method and apparatus, computer device, and storage medium
US20230138424A1 (en) * 2021-10-29 2023-05-04 Zoom Video Communications, Inc. Displaying time zone-specific content in video conferences
US11954710B2 (en) * 2020-01-10 2024-04-09 Tencent Technology (Shenzhen) Company Limited Item display method and apparatus, computer device, and storage medium

Families Citing this family (43)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101496387B (en) 2006-03-06 2012-09-05 思科技术公司 System and method for access authentication in a mobile wireless network
KR101513616B1 (en) * 2007-07-31 2015-04-20 엘지전자 주식회사 Mobile terminal and image information managing method therefor
KR20090032559A (en) * 2007-09-28 2009-04-01 엘지전자 주식회사 Mobile terminal and control method thereof
CN101494681B (en) * 2008-01-22 2012-12-26 宏达国际电子股份有限公司 Information prompting method
US8797377B2 (en) 2008-02-14 2014-08-05 Cisco Technology, Inc. Method and system for videoconference configuration
TWI428814B (en) * 2008-04-15 2014-03-01 Htc Corp Method for switching wallpaper in screen lock state, mobile electronic device thereof, and recording medium thereof
US8390667B2 (en) 2008-04-15 2013-03-05 Cisco Technology, Inc. Pop-up PIP for people not in picture
JP2010016486A (en) * 2008-07-01 2010-01-21 Canon Inc Digital broadcast receiving apparatus and control method and program for the same
US8620294B2 (en) 2008-09-25 2013-12-31 Microsoft Corporation Mobile device dynamic background
KR101517618B1 (en) * 2008-10-31 2015-05-04 삼성전자주식회사 Apparatus and method for displaying web contents effectively
KR20100113266A (en) * 2009-04-13 2010-10-21 삼성전자주식회사 Apparatus and method for manufacturing three-dementions image message in electronic terminal
US9082297B2 (en) 2009-08-11 2015-07-14 Cisco Technology, Inc. System and method for verifying parameters in an audiovisual environment
TW201128516A (en) * 2010-02-12 2011-08-16 Utw Technologies Corp An instinctive operating method and an electronic device using the same
CN101867634A (en) * 2010-03-03 2010-10-20 深圳市同洲电子股份有限公司 Information prompting method and device in mobile terminal
US8896655B2 (en) 2010-08-31 2014-11-25 Cisco Technology, Inc. System and method for providing depth adaptive video conferencing
US8902244B2 (en) 2010-11-15 2014-12-02 Cisco Technology, Inc. System and method for providing enhanced graphics in a video environment
US9338394B2 (en) 2010-11-15 2016-05-10 Cisco Technology, Inc. System and method for providing enhanced audio in a video environment
US8542264B2 (en) 2010-11-18 2013-09-24 Cisco Technology, Inc. System and method for managing optics in a video environment
US8723914B2 (en) 2010-11-19 2014-05-13 Cisco Technology, Inc. System and method for providing enhanced video processing in a network environment
US9111138B2 (en) 2010-11-30 2015-08-18 Cisco Technology, Inc. System and method for gesture interface control
CN102065177A (en) * 2010-12-02 2011-05-18 东莞宇龙通信科技有限公司 Dynamic icon display system and method for mobile terminal
USD682854S1 (en) 2010-12-16 2013-05-21 Cisco Technology, Inc. Display screen for graphical user interface
USD678320S1 (en) * 2010-12-16 2013-03-19 Cisco Technology, Inc. Display screen with graphical user interface
JP2012133659A (en) * 2010-12-22 2012-07-12 Fujifilm Corp File format, server, electronic comic viewer device and electronic comic generation device
US8689116B2 (en) * 2011-01-14 2014-04-01 Apple Inc. Email user interface
JP5929034B2 (en) * 2011-08-09 2016-06-01 ソニー株式会社 Information processing apparatus, information processing method, and program
US10523804B2 (en) 2011-09-16 2019-12-31 Samsung Electronics Co., Ltd. Method and system for searching for object in network
KR101930510B1 (en) * 2011-09-16 2018-12-19 삼성전자주식회사 Method and system for searching object in network
US8817189B2 (en) * 2011-09-29 2014-08-26 Lsi Corporation Digital television with improved input selection functionality
CN102945116A (en) * 2012-10-19 2013-02-27 广东欧珀移动通信有限公司 Interface switching display method and device, and mobile terminal
KR102192159B1 (en) 2013-07-25 2020-12-16 삼성전자주식회사 Method for displaying and an electronic device thereof
CN104898821B (en) * 2014-03-03 2018-01-23 联想(北京)有限公司 The method and electronic equipment of a kind of information processing
KR101652039B1 (en) * 2014-10-30 2016-08-29 주식회사 피티엑스그룹 The method for controlling device by character and aplication
US20160132201A1 (en) * 2014-11-06 2016-05-12 Microsoft Technology Licensing, Llc Contextual tabs in mobile ribbons
US20160307028A1 (en) * 2015-04-16 2016-10-20 Mikhail Fedorov Storing, Capturing, Updating and Displaying Life-Like Models of People, Places And Objects
US20170317960A1 (en) * 2016-04-28 2017-11-02 Jamdeo Canada Ltd. Device and methods for messaging application control and presentation
WO2018172472A1 (en) * 2017-03-22 2018-09-27 Mimi Hearing Technologies GmbH Methods for user-dependent conditioning of stimuli in tests using a method of continuous adjustment
WO2019123744A1 (en) * 2017-12-22 2019-06-27 ソニー株式会社 Information processing device, information processing method, and program
KR20190081561A (en) * 2017-12-29 2019-07-09 삼성전자주식회사 Electronic device and Method for controlling the electronic device thereof
US10630824B2 (en) 2018-03-22 2020-04-21 Apple Inc. Electronic devices with adjustable decoration
US11049169B2 (en) 2018-07-24 2021-06-29 International Business Machines Corporation System, computer program product, and method for automated gift determination and delivery
EP3871195A4 (en) * 2018-12-21 2021-12-22 Samsung Electronics Co., Ltd. Electronic device and method for providing avatar based on emotion state of user
CN113110786B (en) * 2021-06-15 2021-09-03 南京统信软件技术有限公司 Page sliding control method and mobile terminal

Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5682469A (en) * 1994-07-08 1997-10-28 Microsoft Corporation Software platform having a real world interface with animated characters
US20020009988A1 (en) * 2000-07-07 2002-01-24 Kunio Murata Propitious information distribution system using a wireless mobile terminal
US20020198009A1 (en) * 2001-06-26 2002-12-26 Asko Komsi Entity reply mechanism
US6845486B2 (en) * 2000-06-30 2005-01-18 Sanyo Electric Co., Ltd. User support method and user support apparatus
US20050044500A1 (en) * 2003-07-18 2005-02-24 Katsunori Orimoto Agent display device and agent display method
US20050081150A1 (en) * 2001-11-02 2005-04-14 Beardow Paul Rowland Method and apparatus for text messaging
US20050190188A1 (en) * 2004-01-30 2005-09-01 Ntt Docomo, Inc. Portable communication terminal and program
US20050261031A1 (en) * 2004-04-23 2005-11-24 Jeong-Wook Seo Method for displaying status information on a mobile terminal
US20060294465A1 (en) * 2005-06-22 2006-12-28 Comverse, Inc. Method and system for creating and distributing mobile avatars
US7522912B2 (en) * 2004-07-07 2009-04-21 Samsung Electric Co., Ltd. Device and method for downloading character image from website in wireless terminal
US7725842B2 (en) * 2003-04-24 2010-05-25 Bronkema Valentina G Self-attainable analytic tool and method for adaptive behavior modification

Family Cites Families (28)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5347628A (en) * 1990-01-18 1994-09-13 International Business Machines Corporation Method of graphically accessing electronic data
WO1997008879A2 (en) * 1995-08-31 1997-03-06 Philips Electronics N.V. Terminal
JPH1033830A (en) 1996-04-26 1998-02-10 Namco Ltd Game device and information storage medium
DE19749722A1 (en) * 1997-11-11 1999-05-12 Agfa Gevaert Ag Color photographic material
JP3990798B2 (en) 1998-03-09 2007-10-17 キヤノン株式会社 Sheet processing apparatus and image forming apparatus having the same
JP2000049924A (en) * 1998-07-27 2000-02-18 Matsushita Electric Ind Co Ltd Portable telephone set and security method for outgoing and incoming call telephone number
KR100549922B1 (en) * 1999-02-25 2006-02-06 주식회사 팬택앤큐리텔 Method for embodying a telephone directory function of mobile stations
KR20010073430A (en) * 2000-01-14 2001-08-01 황재엽 How to embody the phonebook added the privacy function for handphone
JP2001290490A (en) * 2000-01-31 2001-10-19 Casio Comput Co Ltd Graphic data generating and editing system, digital audio player, graphic data generating and editing method and recording medium
JP2002091971A (en) * 2000-09-11 2002-03-29 Sony Corp Agent system, method/device for providing information and data recording medium
FI115260B (en) * 2000-09-18 2005-03-31 Nokia Corp A method of forming an information element and an electronic device
US6801220B2 (en) * 2001-01-26 2004-10-05 International Business Machines Corporation Method and apparatus for adjusting subpixel intensity values based upon luminance characteristics of the subpixels for improved viewing angle characteristics of liquid crystal displays
JP4096541B2 (en) * 2001-10-01 2008-06-04 株式会社日立製作所 Screen display method
JP2004033690A (en) 2002-07-03 2004-02-05 Square Enix Co Ltd Video game unit, record medium, and program
EP1396984B1 (en) * 2002-09-04 2014-03-05 Hewlett-Packard Development Company, L.P. User interface for a mobile communication device
CN1481185A (en) 2002-09-06 2004-03-10 �����ɷ� Handset capable of displaying custom tailored motion picture and related method
US6903743B2 (en) * 2002-10-16 2005-06-07 Motorola, Inc. Dynamic interactive animated screen saver
KR100932483B1 (en) * 2002-11-20 2009-12-17 엘지전자 주식회사 Mobile communication terminal and avatar remote control method using the same
CN1532652A (en) * 2003-03-19 2004-09-29 泰商泰金宝科技股份有限公司 Method for displaying time
US7224963B2 (en) * 2003-10-17 2007-05-29 Sony Ericsson Mobile Communications Ab System method and computer program product for managing themes in a mobile phone
KR100657065B1 (en) * 2004-01-29 2006-12-13 삼성전자주식회사 Device and method for character processing in wireless terminal
CN1684478A (en) 2004-04-12 2005-10-19 英华达(南京)科技有限公司 Method for increasing schedule set according the lunar calendar in mobile phone
US7359688B2 (en) 2004-04-23 2008-04-15 Samsung Electronics Co., Ltd. Device and method for displaying a status of a portable terminal by using a character image
CN1694463A (en) 2004-04-30 2005-11-09 三星电子株式会社 Method for displaying a screen image on a mobile terminal
US7555717B2 (en) * 2004-04-30 2009-06-30 Samsung Electronics Co., Ltd. Method for displaying screen image on mobile terminal
KR100557130B1 (en) * 2004-05-14 2006-03-03 삼성전자주식회사 Terminal equipment capable of editing movement of avatar and method therefor
US7519924B2 (en) * 2004-11-03 2009-04-14 Research In Motion Limited Handheld electronic device including appointment and meeting conflict notification, and associated method
KR100756731B1 (en) * 2005-05-13 2007-09-07 엘지전자 주식회사 Refregerator with each user's avatar displaying function

Patent Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5682469A (en) * 1994-07-08 1997-10-28 Microsoft Corporation Software platform having a real world interface with animated characters
US6388665B1 (en) * 1994-07-08 2002-05-14 Microsoft Corporation Software platform having a real world interface with animated characters
US6845486B2 (en) * 2000-06-30 2005-01-18 Sanyo Electric Co., Ltd. User support method and user support apparatus
US20020009988A1 (en) * 2000-07-07 2002-01-24 Kunio Murata Propitious information distribution system using a wireless mobile terminal
US20020198009A1 (en) * 2001-06-26 2002-12-26 Asko Komsi Entity reply mechanism
US20050081150A1 (en) * 2001-11-02 2005-04-14 Beardow Paul Rowland Method and apparatus for text messaging
US7725842B2 (en) * 2003-04-24 2010-05-25 Bronkema Valentina G Self-attainable analytic tool and method for adaptive behavior modification
US20050044500A1 (en) * 2003-07-18 2005-02-24 Katsunori Orimoto Agent display device and agent display method
US20050190188A1 (en) * 2004-01-30 2005-09-01 Ntt Docomo, Inc. Portable communication terminal and program
US20050261031A1 (en) * 2004-04-23 2005-11-24 Jeong-Wook Seo Method for displaying status information on a mobile terminal
US7522912B2 (en) * 2004-07-07 2009-04-21 Samsung Electric Co., Ltd. Device and method for downloading character image from website in wireless terminal
US20060294465A1 (en) * 2005-06-22 2006-12-28 Comverse, Inc. Method and system for creating and distributing mobile avatars

Cited By (130)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9665384B2 (en) 2005-08-30 2017-05-30 Microsoft Technology Licensing, Llc Aggregation of computing device settings
US8690664B2 (en) 2006-09-25 2014-04-08 Etasse Limited Slot machine game with additional award indicator
US20080119261A1 (en) * 2006-10-23 2008-05-22 Jorge Heymann Slot machine bonus game providing awards for manual dexterity
US9165419B2 (en) * 2006-10-23 2015-10-20 Etasse Limited Slot machine bonus game providing awards for manual dexterity
US8702493B2 (en) 2007-11-09 2014-04-22 Etasse Limited Slot machine game with award based on another machine
US20090124352A1 (en) * 2007-11-13 2009-05-14 Ignacio Gerson Slot machine game with side pot
US20090199110A1 (en) * 2008-02-05 2009-08-06 Samsung Electronics Co., Ltd. Apparatus and method for transmitting animation-based message
US9520031B2 (en) 2008-07-07 2016-12-13 Etasse Limited Slot machine game with symbol lock-in
US8825699B2 (en) 2008-10-23 2014-09-02 Rovi Corporation Contextual search by a mobile communications device
US8250494B2 (en) 2008-10-23 2012-08-21 Microsoft Corporation User interface with parallax animation
US8086275B2 (en) 2008-10-23 2011-12-27 Microsoft Corporation Alternative inputs of a mobile communications device
US20100107068A1 (en) * 2008-10-23 2010-04-29 Butcher Larry R User Interface with Parallax Animation
US8781533B2 (en) 2008-10-23 2014-07-15 Microsoft Corporation Alternative inputs of a mobile communications device
US9218067B2 (en) 2008-10-23 2015-12-22 Microsoft Technology Licensing, Llc Mobile communications device user interface
US10133453B2 (en) 2008-10-23 2018-11-20 Microsoft Technology Licensing, Llc Alternative inputs of a mobile communications device
US9323424B2 (en) 2008-10-23 2016-04-26 Microsoft Corporation Column organization of content
US20100105424A1 (en) * 2008-10-23 2010-04-29 Smuga Michael A Mobile Communications Device User Interface
US9703452B2 (en) 2008-10-23 2017-07-11 Microsoft Technology Licensing, Llc Mobile communications device user interface
US8970499B2 (en) 2008-10-23 2015-03-03 Microsoft Technology Licensing, Llc Alternative inputs of a mobile communications device
US9606704B2 (en) 2008-10-23 2017-03-28 Microsoft Technology Licensing, Llc Alternative inputs of a mobile communications device
US8385952B2 (en) 2008-10-23 2013-02-26 Microsoft Corporation Mobile communications device user interface
US8411046B2 (en) 2008-10-23 2013-04-02 Microsoft Corporation Column organization of content
US9223412B2 (en) 2008-10-23 2015-12-29 Rovi Technologies Corporation Location-based display characteristics in a user interface
US8634876B2 (en) 2008-10-23 2014-01-21 Microsoft Corporation Location based display characteristics in a user interface
US9223411B2 (en) 2008-10-23 2015-12-29 Microsoft Technology Licensing, Llc User interface with parallax animation
US8428658B2 (en) * 2009-03-02 2013-04-23 Lg Electronics Inc. Method for displaying items in mobile terminal and mobile terminal using the same
US20100222112A1 (en) * 2009-03-02 2010-09-02 Seung Sook Han Method for displaying items in mobile terminal and mobile terminal using the same
US8548431B2 (en) 2009-03-30 2013-10-01 Microsoft Corporation Notifications
WO2010117642A3 (en) * 2009-03-30 2011-01-13 Microsoft Corporation Unlock screen
US20100248688A1 (en) * 2009-03-30 2010-09-30 Teng Stephanie E Notifications
US8892170B2 (en) 2009-03-30 2014-11-18 Microsoft Corporation Unlock screen
US8355698B2 (en) 2009-03-30 2013-01-15 Microsoft Corporation Unlock screen
US20100248787A1 (en) * 2009-03-30 2010-09-30 Smuga Michael A Chromeless User Interface
US8238876B2 (en) 2009-03-30 2012-08-07 Microsoft Corporation Notifications
US9977575B2 (en) 2009-03-30 2018-05-22 Microsoft Technology Licensing, Llc Chromeless user interface
US8175653B2 (en) 2009-03-30 2012-05-08 Microsoft Corporation Chromeless user interface
US8914072B2 (en) 2009-03-30 2014-12-16 Microsoft Corporation Chromeless user interface
US20120026174A1 (en) * 2009-04-27 2012-02-02 Sonoma Data Solution, Llc Method and Apparatus for Character Animation
US8269736B2 (en) 2009-05-22 2012-09-18 Microsoft Corporation Drop target gestures
US8836648B2 (en) 2009-05-27 2014-09-16 Microsoft Corporation Touch pull-in gesture
US9372701B2 (en) * 2010-05-12 2016-06-21 Sony Interactive Entertainment America Llc Management of digital information via a buoyant interface moving in three-dimensional space
US20110283238A1 (en) * 2010-05-12 2011-11-17 George Weising Management of Digital Information via an Interface
US20130088516A1 (en) * 2010-05-17 2013-04-11 Ntt Docomo, Inc. Object displaying apparatus, object displaying system, and object displaying method
US9449308B2 (en) * 2010-12-14 2016-09-20 Microsoft Technology Licensing, Llc Defining actions for data streams via icons
US20120151381A1 (en) * 2010-12-14 2012-06-14 Microsoft Corporation Defining actions for data streams via icons
US9430130B2 (en) 2010-12-20 2016-08-30 Microsoft Technology Licensing, Llc Customization of an immersive environment
US9696888B2 (en) 2010-12-20 2017-07-04 Microsoft Technology Licensing, Llc Application-launching interface for multiple modes
US8990733B2 (en) 2010-12-20 2015-03-24 Microsoft Technology Licensing, Llc Application-launching interface for multiple modes
US9870132B2 (en) 2010-12-23 2018-01-16 Microsoft Technology Licensing, Llc Application reporting in an application-selectable user interface
US8689123B2 (en) 2010-12-23 2014-04-01 Microsoft Corporation Application reporting in an application-selectable user interface
US9229918B2 (en) 2010-12-23 2016-01-05 Microsoft Technology Licensing, Llc Presenting an application change through a tile
US8560959B2 (en) 2010-12-23 2013-10-15 Microsoft Corporation Presenting an application change through a tile
US9864494B2 (en) 2010-12-23 2018-01-09 Microsoft Technology Licensing, Llc Application reporting in an application-selectable user interface
US11126333B2 (en) 2010-12-23 2021-09-21 Microsoft Technology Licensing, Llc Application reporting in an application-selectable user interface
US9766790B2 (en) 2010-12-23 2017-09-19 Microsoft Technology Licensing, Llc Application reporting in an application-selectable user interface
US8612874B2 (en) 2010-12-23 2013-12-17 Microsoft Corporation Presenting an application change through a tile
US9015606B2 (en) 2010-12-23 2015-04-21 Microsoft Technology Licensing, Llc Presenting an application change through a tile
US9213468B2 (en) 2010-12-23 2015-12-15 Microsoft Technology Licensing, Llc Application reporting in an application-selectable user interface
US10969944B2 (en) 2010-12-23 2021-04-06 Microsoft Technology Licensing, Llc Application reporting in an application-selectable user interface
US20120165071A1 (en) * 2010-12-28 2012-06-28 Inventec Appliances (Shanghai) Co. Ltd. Mobile device capable of automatically switching its operation modes
US9423951B2 (en) 2010-12-31 2016-08-23 Microsoft Technology Licensing, Llc Content-based snap point
US20130201204A1 (en) * 2011-01-27 2013-08-08 Tencent Technology (Shenzhen) Company Limited Method And Device For Implementing A Generic Icon With Multiple Display Modes
CN102622214A (en) * 2011-01-27 2012-08-01 腾讯科技(深圳)有限公司 Method and device for realizing multiple-display mode universal icons
US9383917B2 (en) 2011-03-28 2016-07-05 Microsoft Technology Licensing, Llc Predictive tiling
US9158445B2 (en) 2011-05-27 2015-10-13 Microsoft Technology Licensing, Llc Managing an immersive interface in a multi-application immersive environment
US9052820B2 (en) 2011-05-27 2015-06-09 Microsoft Technology Licensing, Llc Multi-application environment
US11272017B2 (en) 2011-05-27 2022-03-08 Microsoft Technology Licensing, Llc Application notifications manifest
US9658766B2 (en) 2011-05-27 2017-05-23 Microsoft Technology Licensing, Llc Edge gesture
US9329774B2 (en) 2011-05-27 2016-05-03 Microsoft Technology Licensing, Llc Switching back to a previously-interacted-with application
US8893033B2 (en) 2011-05-27 2014-11-18 Microsoft Corporation Application notifications
US11698721B2 (en) 2011-05-27 2023-07-11 Microsoft Technology Licensing, Llc Managing an immersive interface in a multi-application immersive environment
US9104307B2 (en) 2011-05-27 2015-08-11 Microsoft Technology Licensing, Llc Multi-application environment
US9104440B2 (en) 2011-05-27 2015-08-11 Microsoft Technology Licensing, Llc Multi-application environment
US9535597B2 (en) 2011-05-27 2017-01-03 Microsoft Technology Licensing, Llc Managing an immersive interface in a multi-application immersive environment
US10303325B2 (en) 2011-05-27 2019-05-28 Microsoft Technology Licensing, Llc Multi-application environment
US8687023B2 (en) 2011-08-02 2014-04-01 Microsoft Corporation Cross-slide gesture to select and rearrange
US10579250B2 (en) 2011-09-01 2020-03-03 Microsoft Technology Licensing, Llc Arranging tiles
US8935631B2 (en) 2011-09-01 2015-01-13 Microsoft Corporation Arranging tiles
US8922575B2 (en) 2011-09-09 2014-12-30 Microsoft Corporation Tile cache
US9557909B2 (en) 2011-09-09 2017-01-31 Microsoft Technology Licensing, Llc Semantic zoom linguistic helpers
US10353566B2 (en) 2011-09-09 2019-07-16 Microsoft Technology Licensing, Llc Semantic zoom animations
US10114865B2 (en) 2011-09-09 2018-10-30 Microsoft Technology Licensing, Llc Tile cache
US10254955B2 (en) 2011-09-10 2019-04-09 Microsoft Technology Licensing, Llc Progressively indicating new content in an application-selectable user interface
US8830270B2 (en) 2011-09-10 2014-09-09 Microsoft Corporation Progressively indicating new content in an application-selectable user interface
US9146670B2 (en) 2011-09-10 2015-09-29 Microsoft Technology Licensing, Llc Progressively indicating new content in an application-selectable user interface
US9244802B2 (en) 2011-09-10 2016-01-26 Microsoft Technology Licensing, Llc Resource user interface
US8933952B2 (en) 2011-09-10 2015-01-13 Microsoft Corporation Pre-rendering new content for an application-selectable user interface
US9817544B2 (en) * 2011-11-28 2017-11-14 Kyocera Corporation Device, method, and storage medium storing program
US20150370448A1 (en) * 2011-11-28 2015-12-24 Kyocera Corporation Device, method, and storage medium storing program
US9223472B2 (en) 2011-12-22 2015-12-29 Microsoft Technology Licensing, Llc Closing applications
US10191633B2 (en) 2011-12-22 2019-01-29 Microsoft Technology Licensing, Llc Closing applications
US9128605B2 (en) 2012-02-16 2015-09-08 Microsoft Technology Licensing, Llc Thumbnail-image selection of applications
US20130262511A1 (en) * 2012-04-02 2013-10-03 Google Inc. Determining 3D Model Information From Stored Images
US8976179B2 (en) * 2012-04-02 2015-03-10 Google Inc. Determining 3D model information from stored images
US20180024714A1 (en) * 2012-04-04 2018-01-25 Samsung Electronics Co., Ltd. Intelligent event information presentation method and terminal
US10827306B2 (en) * 2012-04-04 2020-11-03 Samsung Electronics Co., Ltd Intelligent event information presentation method and terminal
US9316505B2 (en) * 2013-01-10 2016-04-19 Fujitsu Limited Analysis method, and analysis apparatus
US20140195144A1 (en) * 2013-01-10 2014-07-10 Fujitsu Limited Analysis method, and analysis apparatus
US9807081B2 (en) 2013-05-29 2017-10-31 Microsoft Technology Licensing, Llc Live tiles without application-code execution
US10110590B2 (en) 2013-05-29 2018-10-23 Microsoft Technology Licensing, Llc Live tiles without application-code execution
US9450952B2 (en) 2013-05-29 2016-09-20 Microsoft Technology Licensing, Llc Live tiles without application-code execution
US10531852B2 (en) * 2013-12-24 2020-01-14 Samsung Electronics Co., Ltd. Display for converting medical images and method of use
US20150278765A1 (en) * 2014-04-01 2015-10-01 Microsoft Corporation Information collections
US9841874B2 (en) 2014-04-04 2017-12-12 Microsoft Technology Licensing, Llc Expandable application representation
US10459607B2 (en) 2014-04-04 2019-10-29 Microsoft Technology Licensing, Llc Expandable application representation
US9769293B2 (en) 2014-04-10 2017-09-19 Microsoft Technology Licensing, Llc Slider cover for computing device
US9451822B2 (en) 2014-04-10 2016-09-27 Microsoft Technology Licensing, Llc Collapsible shell cover for computing device
US10685186B2 (en) * 2014-06-06 2020-06-16 Beijing Sogou Technology Development Co., Ltd. Semantic understanding based emoji input method and device
US20170052946A1 (en) * 2014-06-06 2017-02-23 Siyu Gu Semantic understanding based emoji input method and device
US10254942B2 (en) 2014-07-31 2019-04-09 Microsoft Technology Licensing, Llc Adaptive sizing and positioning of application windows
US10592080B2 (en) 2014-07-31 2020-03-17 Microsoft Technology Licensing, Llc Assisted presentation of application windows
US10678412B2 (en) 2014-07-31 2020-06-09 Microsoft Technology Licensing, Llc Dynamic joint dividers for application windows
US10642365B2 (en) 2014-09-09 2020-05-05 Microsoft Technology Licensing, Llc Parametric inertia and APIs
US20160080298A1 (en) * 2014-09-12 2016-03-17 Samsung Electronics Co., Ltd. Method for generating emoticon and electronic device supporting the same
US10068256B2 (en) 2014-10-08 2018-09-04 Microsoft Technology Licensing, Llc User directed information collections
US20160125715A1 (en) * 2014-10-30 2016-05-05 Fih (Hong Kong) Limited Electronic device and alarm method of the electronic device
US9674335B2 (en) 2014-10-30 2017-06-06 Microsoft Technology Licensing, Llc Multi-configuration input device
US10274903B2 (en) 2015-03-27 2019-04-30 Alibaba Group Holding Limited Setting an alarm clock on a smart device
US9684284B2 (en) 2015-03-27 2017-06-20 Alibaba Group Holding Limited Setting an alarm clock on a smart device
US20160352667A1 (en) * 2015-06-01 2016-12-01 Facebook, Inc. Providing augmented message elements in electronic communication threads
US10791081B2 (en) 2015-06-01 2020-09-29 Facebook, Inc. Providing augmented message elements in electronic communication threads
US11233762B2 (en) 2015-06-01 2022-01-25 Facebook, Inc. Providing augmented message elements in electronic communication threads
US10225220B2 (en) * 2015-06-01 2019-03-05 Facebook, Inc. Providing augmented message elements in electronic communication threads
US11350007B2 (en) * 2018-12-07 2022-05-31 Canon Kabushiki Kaisha Image reading apparatus, image processing method, and storage medium for correcting show-through
WO2020128962A1 (en) * 2018-12-19 2020-06-25 RxPrism Health Systems Pvt. Ltd A system and a method for creating and sharing interactive content rapidly anywhere and anytime
JP2021139904A (en) * 2019-12-24 2021-09-16 株式会社ユピテル Device and program
JP7209385B2 (en) 2019-12-24 2023-01-20 株式会社ユピテル Device and program
US20220245681A1 (en) * 2020-01-10 2022-08-04 Tencent Technology (Shenzhen) Company Limited Item display method and apparatus, computer device, and storage medium
US11954710B2 (en) * 2020-01-10 2024-04-09 Tencent Technology (Shenzhen) Company Limited Item display method and apparatus, computer device, and storage medium
US20230138424A1 (en) * 2021-10-29 2023-05-04 Zoom Video Communications, Inc. Displaying time zone-specific content in video conferences

Also Published As

Publication number Publication date
EP1796356A2 (en) 2007-06-13
EP1796356A3 (en) 2013-01-23
EP1796355A3 (en) 2013-05-29
US8132100B2 (en) 2012-03-06
CN1980268B (en) 2010-05-26
CN1980441B (en) 2010-05-26
US20070183381A1 (en) 2007-08-09
KR20080084908A (en) 2008-09-22
KR100785067B1 (en) 2007-12-12
KR20070059951A (en) 2007-06-12
CN1980268A (en) 2007-06-13
CN1980441A (en) 2007-06-13
EP1796355A2 (en) 2007-06-13
KR20070059950A (en) 2007-06-12

Similar Documents

Publication Publication Date Title
US20070171192A1 (en) Screen image presentation apparatus and method for mobile phone
US11849255B2 (en) Multi-participant live communication user interface
JP6842799B2 (en) Sharing user-configurable graphic structures
US8111267B2 (en) Controlling image and mobile terminal
US9372701B2 (en) Management of digital information via a buoyant interface moving in three-dimensional space
CN107066192A (en) Equipment, method and graphic user interface for manipulating user interface object using vision and/or touch feedback
JP6960513B2 (en) Creative camera
CN110471581A (en) The user interface that equipment for controlling or presenting on electronic equipment uses
CN102687105B (en) The system and method for configurable user interface controller while being used for moving
CN109766045A (en) Method and apparatus for the operating function in touching device
CN105103111A (en) User interface for computing device
CN107820116A (en) Video broadcasting method and device
KR20230078606A (en) User interfaces related to time
CN115378893A (en) Message processing method and device, electronic equipment and readable storage medium
US11868601B2 (en) Devices, methods, and graphical user interfaces for providing notifications and application information
KR102265402B1 (en) Multi-Participant Live Communication User Interface
CN114764361A (en) Expression special effect display method, device, terminal and storage medium
Toř A Discourse Analysis of Important Apple Keynotes

Legal Events

Date Code Title Description
AS Assignment

Owner name: SAMSUNG ELECTRONICS CO., LTD., KOREA, REPUBLIC OF

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:SEO, JEONG WOOK;LEE, JONG KERL;PARK, WEI JIN;REEL/FRAME:019143/0426

Effective date: 20070305

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION