CN112954209A - Photographing method and device, electronic equipment and medium - Google Patents
- ️Fri Jun 11 2021
Detailed Description
The technical solutions in the embodiments of the present application will be clearly and completely described below with reference to the drawings in the embodiments of the present application, and it is obvious that the described embodiments are some, but not all, embodiments of the present application. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present application.
The terms first, second and the like in the description and in the claims of the present application are used for distinguishing between similar elements and not necessarily for describing a particular sequential or chronological order. It will be appreciated that the data so used may be interchanged under appropriate circumstances such that embodiments of the application may be practiced in sequences other than those illustrated or described herein, and that the terms "first," "second," and the like are generally used herein in a generic sense and do not limit the number of terms, e.g., the first term can be one or more than one. In addition, "and/or" in the specification and claims means at least one of connected objects, a character "/" generally means that a preceding and succeeding related objects are in an "or" relationship.
The photographing method provided by the embodiment of the present application will be described in detail by specific embodiments and application scenarios thereof with reference to the accompanying drawings.
Fig. 1 is a schematic flow chart of a photographing method according to an embodiment of the present disclosure. The implementation of the method is described in detail below with reference to this figure.
101, acquiring first motion information of eyeballs of a user under the condition that a shooting preview interface is displayed, wherein the shooting preview interface comprises N preview sub-areas, each preview sub-area is associated with a camera, and N is a positive integer;
in this step, the electronic device is triggered to enter a shooting preview interface when receiving a shooting request.
Here, the shooting preview interface may be divided into N preview sub-areas by a preset input of the user to the shooting preview interface. Here, the preset input is used to trigger the electronic device to enter the first photographing mode. It should be noted that the first photographing mode is a photographing mode in which a camera used for photographing is determined by eye movement of a user.
It should be noted that each preview sub-area is associated with one different camera corresponding to different preview sub-areas.
Here, each preview sub-area displays a preview image based on a picture captured by the corresponding camera. The preview images displayed in different preview sub-areas are different, so that a user can conveniently and visually check the imaging effects of different cameras of the electronic equipment.
102, determining target preview sub-areas watched by eyeballs of the user based on the first motion information, wherein the N preview sub-areas comprise the target preview sub-areas;
103, controlling a target camera associated with the target preview sub-area to shoot a target image under the condition that the time length of the eyeball of the user watching the target preview sub-area is within a preset time length range.
In this step, in order to avoid that the gaze of the eyeball of the user on the target preview subregion is invalid gaze, under the condition that the time length of the eyeball of the user gazing on the target preview subregion is within the preset time length range, the camera associated with the target preview subregion is determined to be the target camera, that is, the target camera is the camera selected by the user and finally used for photographing; and controls the target camera to shoot the target image.
For example, the preset time duration range is 3 seconds to 5 seconds, and the time duration when the eyeballs of the user watch the target preview sub-area is less than 3 seconds or more than 5 seconds, the user is judged to watch the target preview sub-area inefficiently. When the time length that the eyeballs of the user watch the target preview sub-area is less than 3 seconds, the eyeballs of the user watch the target preview sub-area for a short time, a large possibility that the subsequent sight line leaves the area is provided, when the time length that the eyeballs of the user watch the target preview sub-area is more than 5 seconds, the user is considered to be in a dull state or meditation, and the camera which is finally used for shooting and is wanted by the user is not provided, and the above situations are considered to be invalid watching.
According to the photographing method, first motion information of eyeballs of a user is acquired under the condition that a photographing preview interface is displayed, wherein the photographing preview interface comprises N preview sub-areas, each preview sub-area is associated with one camera, and N is a positive integer; determining a target preview subregion watched by the eyeballs of the user based on the first motion information, wherein the N preview subregions comprise the target preview subregion, and the N preview subregions comprise the target preview subregion; under the condition that the time length of the target preview subregion is watched by eyeballs of the user, the target camera associated with the target preview subregion is controlled to shoot a target image, so that the camera required by the user can be conveniently and quickly selected to shoot through the movement of the eyeballs of the user, the shooting mode is more flexible, the camera is particularly suitable for scenes in which the hands of the user are inconvenient to operate, manual operation is not needed, and the required camera can be flexibly and quickly selected to shoot the target image by controlling the movement of the eyeballs to select one region.
Optionally, each of the N preview sub-areas includes an area identifier, and the area identifier has a first display mode and a second display mode;
under the condition that the area identification is in the first display mode, the area identification is used for indicating the eyeball of the user to watch the preview sub-area where the area identification is located;
and under the condition that the area identification is in the second display mode, the area identification is used for indicating that eyeballs of the user do not watch the preview sub-area where the area identification is located.
Here, the area identification of each preview sub-area is used to indicate whether the current user's eye looks at the preview sub-area.
In one example, through user operation, the electronic device opens a camera application and enters a shooting preview interface; displaying 6 preview sub-areas on a shooting preview interface through preset input of a user on the shooting preview interface, as shown in fig. 2; wherein the lower right corner of each preview sub-area displays an eye-style identification (i.e. a region identification). When the eyeball of the user does not watch the shooting preview interface, the identification of the eye pattern of each preview subarea is a second display mode 1, such as being grayed; when the user's eyes are focused on one of the 6 preview sub-areas, the eye pattern of the preview sub-area is identified as the
first display mode2, such as presenting a preset color.
Based on this, after determining, in
step102, a target preview sub-area at which the user looks at based on the first motion information, an embodiment of the present application may further include:
and updating the area identification in the target preview subregion to the first display mode.
Continuing with the above example illustration, that is, when the user looks at the target preview sub-area, the area identifier of the target preview sub-area, such as the eye style identifier, is updated to present the preset color from the original grayed-out. Therefore, the user can intuitively know the preview subarea watched by the eyeballs in real time.
As an optional implementation manner, the shooting preview interface further includes a position identifier, where the position identifier is used to indicate first motion information of the eyeballs of the user, and the first motion information includes a real-time position watched by the eyeballs of the user;
based on this, the
step102 of determining, based on the first motion information, a target preview sub-area at which the user looks at the eyeball may include:
in a target time period, under the condition that the display position of the position mark is located in the display range of a first preview subregion, determining the first preview subregion as a target preview subregion watched by eyeballs of a user;
wherein the first preview sub-area is any one of the N preview sub-areas.
It should be noted that, when the user looks at the shooting preview interface, the position identifier moves on the shooting preview interface along with the movement of the user's eyeball, so that the position where the user looks at on the screen, that is, the real-time position where the user's eyeball looks at, can be visually displayed.
Based on this, through the position identification, the motion information of the eyeballs of the user, namely the real-time position watched by the eyeballs of the user, can be obtained, so that the target preview sub-area watched by the target eyeballs is determined, and the processing efficiency of determining the target preview sub-area watched by the target eyeballs can be improved.
Of course, the first motion information of the eyeball of the user can also be acquired through the front camera of the electronic equipment. Wherein the first motion information includes but is not limited to: the offset direction of the user's eyeball and the offset angle of the user's eyeball.
Here, in the case where the photographing preview interface displays N regions, the front camera is turned on, and first movement information of the user's eyes is acquired in real time through the front camera. And calculating to obtain a target preview subregion watched by eyeballs of the user according to the first motion information.
In an example, as shown in fig. 3, in a case where the shooting
preview interface displays6 preview sub-areas, if the user looks at the shooting preview interface, the shooting preview interface displays an identifier 3 (i.e. a position identifier) of a semi-transparent eye pattern at a position of an area where the user's eyes watch, where a dotted line in the figure represents the identifier of the semi-transparent eye pattern. For example, in fig. 3, the current region where the eyes of the user gaze is in the
preview sub-region4 in the middle of the left side, and the translucent eye-pattern mark 3 is displayed in the
preview sub-region4. If the display position of the semi-transparent eye-style identifier 3 is within the display range of the
preview sub-area4 in the middle of the left side in fig. 3 in the target time period, the
preview sub-area4 is determined as a target preview sub-area watched by the eyeballs of the user.
How to display the N preview sub-regions on the photo preview interface is implemented, as an optional implementation manner, the embodiment of the present application may further include:
receiving preset input of a user to a shooting preview interface;
here, the preset input may include, but is not limited to, at least one of a click input, a press input, a long press input, a pinch input, a drag input, a slide input, and a swipe input, that is, the preset input may be one of the above-described inputs, or may also be a combination input of two or more thereof.
And responding to the preset input, and displaying N preview sub-areas on the shooting preview interface.
Optionally, the camera associated with the preview sub-area is a rear camera. The number of the cameras in the realization mode is equal to the number of the preview subareas displayed in the shooting preview interface.
In one example, through user operation, the electronic device opens a camera application and enters a shooting preview interface; the shooting preview interface is slid downwards by two fingers of a user to enter a first shooting mode, namely all rear cameras of the electronic equipment are opened, and if the number of the rear cameras of the electronic equipment is 6, 6 preview sub-areas are displayed on the shooting preview interface, which is shown in fig. 2. Namely, the shooting preview interface is divided into 6 corresponding preview sub-areas according to the number of the rear cameras, each preview sub-area is associated with one rear camera, the cameras associated with different preview sub-areas are different, and each preview sub-area displays image data collected by the associated camera. In this way, the user can preview the image data of 6 cameras at the same time.
As an optional implementation manner, in the case of displaying a shooting preview interface, an embodiment of the present application may further include:
receiving first input of a user to M preview sub-areas in the N preview sub-areas;
here, the first input is a preset input, and the first input may include, but is not limited to, at least one of an eye control input/click input, a press input, a long press input, a pinch input, a drag input, a slide input, and a swipe input, that is, the first input may be one of the above-mentioned inputs, or may also be a combination input of two or more thereof. Here, the eye control input is controlled by the movement of the user's eyeball.
In response to the first input, combining the M preview sub-areas into a second preview sub-area, performing image synthesis on preview images acquired by the M cameras displayed in the M preview sub-areas to generate a target preview image, and displaying the target preview image in the second preview sub-area; wherein M is an integer greater than 1.
The combined area, namely the target preview image displayed in the second preview subregion is an image generated by synthesizing preview images acquired by M cameras displayed by the M preview subregions, so that not only can the combination of any two preview subregions be controlled conveniently, but also a user can directly view multi-shot preview images after image synthesis processing, and the user has a larger selection space for the camera used in the final shooting, thereby realizing more flexible and convenient shooting and meeting the requirements of the user.
In one example, through user operation, the electronic device opens a camera application and enters a shooting preview interface; the shooting preview interface is slid downwards by two fingers of a user to enter a first shooting mode, namely all rear cameras of the electronic equipment are opened, and if the number of the rear cameras of the electronic equipment is 6, 6 preview sub-areas are displayed on the shooting preview interface, which is shown in fig. 2. Namely, the shooting preview interface is divided into 6 corresponding preview sub-areas according to the number of the rear cameras, each preview sub-area is associated with one rear camera, the cameras associated with different preview sub-areas are different, and each preview sub-area displays image data acquired by the associated camera.
Then, referring to the left diagram of fig. 4, a user's slide input (slide to the right in fig. 4) to the shooting preview interface is received, the trajectory of the slide input crossing the two
preview sub-areas5 and 6; in response to the slide input, the two
preview sub-areas5 and 6 are merged into one preview sub-area 7, see the right hand diagram of fig. 4. At this time, the number of preview sub-areas displayed on the shooting preview interface is updated from the original 6 to 5. At this time, the number of cameras associated with the combined preview sub-area 7 is two. Then, the preview images acquired by the 2 cameras displayed in the two
preview sub-areas5 and 6 are synthesized to generate a target preview image, and the target preview image is displayed in the combined preview sub-area 7.
Based on this, if the user wants to restore the original split state again, as an optional implementation manner, in response to the first input, the M preview sub-areas are merged into the second preview sub-area, and the preview images of the M preview sub-areas are image-synthesized to generate a target preview image, and after the target preview image is displayed in the second preview sub-area, the embodiment of the present application may further include:
receiving a second input of the user to the second preview subarea;
here, the second input is a preset input, and the second input may include, but is not limited to, at least one of an eye control input/click input, a press input, a long press input, a pinch input, a drag input, a slide input, and a swipe input, that is, the second input may be one of the above-mentioned inputs, or may also be a combination input of two or more thereof. Here, the eye control input is controlled by the movement of the user's eyeball.
Responding to the second input, splitting the second preview subregion into M preview subregions, and respectively displaying a preview image acquired by each camera in the M cameras in each preview subregion in the M preview subregions.
Here, each preview sub-area of the split M preview sub-areas is associated with one camera, that is, each preview sub-area displays a preview image acquired by one camera.
By a second input, such as a reverse input (leftward sliding input) of the sliding input when merging the regions in the previous example, specifically, the identification of the eye pattern of the merging region is clicked and released after sliding a distance leftward.
Optionally, under the condition that a preview image displayed by a third preview sub-area of the N preview sub-areas is obtained by synthesizing preview images acquired by at least two cameras, the third preview sub-area further includes a camera number identifier, and the camera number identifier is used for indicating the number of cameras associated with the third preview sub-area.
Here, when the number of the cameras associated with the preview sub-area exceeds two, the preview sub-area, here, referred to as a third preview sub-area, includes a camera number identifier for indicating the number of the cameras associated with the third preview sub-area, so that a user can intuitively know the number of the cameras associated with the preview sub-area in a shooting preview interface.
Specifically, referring to the right side of fig. 4, after the two
preview sub-areas5 and 6 are merged into one preview sub-area 7 in response to a sliding input of the user to the shooting preview interface, a
camera number identifier2 is displayed at the lower right corner of the merged preview sub-area 7, so as to indicate that the number of cameras associated with the preview sub-area is 2.
As an optional implementation manner, in a case that a duration of time that the user's eyeball focuses on the target preview sub-area exceeds a preset first duration, the method of the embodiment of the present application may further include:
under the condition of acquiring voice information input by a user, determining a camera obtained by recognizing the voice information as a target camera; or,
under the condition that selection input of a user on a fourth preview subarea in the N preview subareas is received, determining a camera related to the fourth preview subarea as a target camera;
and the preset first time length is greater than or equal to the maximum value of the preset time length range.
Here, when the duration of the target sub-preview area watched by the eyes of the user exceeds a preset first duration, that is, the gaze of the eyes of the user is invalid, the selected result of the camera finally used for photographing is selected as invalid through the movement information of the eyes of the user.
In this scenario, one of the cameras starts a voice recognition function by default to acquire voice information input by a user, and determines a camera obtained by recognizing the voice information as a target camera when the voice information input by the user is acquired. That is, the target camera, which is the camera ultimately used for taking a picture, can be selected through the voice command.
In an example, referring to the left diagram of fig. 5, if the time length of the user's eyeball gazing at the middle left preview sub-area 8 exceeds 5 seconds or longer, it is determined that the user's eyeball gazing is invalid gazing, the voice recognition function is started by default, and the user can directly select one camera through voice recognition, for example, the user says: the "wide-angle camera" or, if the camera associated with the preview sub-area is numbered, the user may also say that the number of the camera is, for example, "
number2", at this time, the target camera, that is, the camera finally used for taking a picture, is determined. Further, referring to the middle diagram of fig. 5, the identifier of the eye pattern of the preview sub-area 9 corresponding to the wide-angle camera is presented with a preset color, and the
identifier10 of the eye pattern of the left area 8 in the middle of the shooting preview interface is grayed out.
In addition, after the camera identified by the user voice information is determined as the target camera (i.e. the camera finally used for photographing is selected by the user voice instruction), a photographing instruction is not received within a preset time period, and updated motion information of the user eyeballs is obtained within the preset time period (i.e. the user eye sight line starts to change and does not look at a place any more) in the preset time period (corresponding to the right side diagram of fig. 5), the method returns to the step after the
step101 of continuing to execute, that is, the camera finally used for photographing is determined by the motion information of the user eyeballs.
And the target camera, namely the camera finally used for photographing is determined through the selection input of the user to the preview subarea.
As an optional implementation manner, before controlling the target camera associated with the target preview sub-area to shoot the target image, the method in this embodiment of the application may further include:
displaying a shooting parameter setting window under the condition that a third input of the user to the target preview sub-area is received, wherein the shooting parameter setting window comprises at least one shooting parameter;
optionally, in a case where a third input of the area identification of the target preview sub-area by the user is received, the shooting parameter setting window is displayed.
Receiving a fourth input of the shooting parameter setting window by the user;
here, in a case where the shooting parameter setting window includes an editing area in which the user can edit the shooting parameters, the fourth input is an editing input for the shooting parameters. That is, by the user manually editing the input, the photographing parameters desired by the user are determined.
In a case where the shooting parameter setting window includes a shooting parameter selection item selectable by the user, the fourth input is a selection input of the target shooting parameter selection item. That is, through the selection input of the user, the photographing parameters desired by the user are determined.
Determining a first photographing parameter in response to the fourth input;
the controlling the target camera associated with the target preview sub-area to shoot the target image comprises the following steps:
and controlling a target camera associated with the target preview subregion to shoot a target image according to the first shooting parameter.
In the implementation mode, the shooting parameters meeting the user requirements can be set through the shooting parameter setting window, and the target image can be shot by adopting the target camera according to the shooting parameters meeting the user requirements, so that the target image meeting the shooting effect desired by the user can be obtained.
In one example, through user operation, the electronic device opens a camera application and enters a shooting preview interface; entering a first photographing mode through the downslide input of two fingers of a user on a photographing preview interface, namely opening all rear cameras of the electronic equipment, and displaying 6 preview sub-areas on the photographing preview interface if the number of the rear cameras of the electronic equipment is 6, referring to fig. 6, wherein an eye-style identifier (namely an area identifier) is displayed at the lower right corner of each preview sub-area; when the eyeball of the user is focused on the right area 10 (target preview sub-area) in the middle of the 6 preview sub-areas, the identification of the eye style of the preview sub-area 10 is a first display mode 2, for example, a preset color is presented, and the time length for the eyeball of the user to watch the preview sub-area 10 is within a preset time length range, it is determined that the camera associated with the preview sub-area 10 is the target camera; then, clicking the mark of the eye pattern of the preview subregion 10, popping up a shooting parameter setting window 11, and selecting shooting time delay and shooting number as a first shooting parameter through user operation; or, the shooting parameter setting window, the subsequent shooting time delay selection and the number of the shot can be realized through eye control input; and finally, clicking a photographing button, and photographing a target image by using a target camera according to the first photographing parameter.
As an optional implementation manner, before controlling the target camera associated with the target preview sub-area to shoot the target image, the embodiment of the present application may further include:
acquiring second motion information of eyeballs of the user;
controlling the position mark to move according to the second motion information;
determining a second shooting parameter according to the moving distance of the position identifier;
that is to say, through the removal of eye control position sign to confirm the second shooting parameter, like this, saved the step that the user edited the selection shooting parameter manually, for the user provides convenience when inconvenient use hand and interface are interacted.
Correspondingly, the controlling the target camera associated with the target preview sub-area to shoot the target image includes:
and controlling the target camera associated with the target preview subarea to shoot the target image according to the second shooting parameter.
According to the photographing method, first motion information of eyeballs of a user is acquired under the condition that a photographing preview interface is displayed, wherein the photographing preview interface comprises N preview sub-areas, each preview sub-area is associated with one camera, and N is a positive integer; determining a target preview subregion watched by the eyeballs of the user based on the first motion information, wherein the N preview subregions comprise the target preview subregion, and the N preview subregions comprise the target preview subregion; under the condition that the time length of the target preview subarea watched by the eyeballs of the user is within the preset time length range, the target camera associated with the target preview subarea is controlled to shoot a target image, so that the camera needed by the user can be conveniently and quickly selected to shoot through the movement of the eyeballs of the user, and the shooting mode is more flexible.
It should be noted that, in the photographing method provided in the embodiment of the present application, the execution main body may be a photographing device, or a control module used for executing the photographing method in the photographing device. The embodiment of the present application takes a photographing apparatus executing a photographing method as an example, and the photographing apparatus provided in the embodiment of the present application is described.
Fig. 7 is a schematic structural diagram of a photographing device according to an embodiment of the present application. The photographing apparatus 700 may include:
a first obtaining
module701, configured to obtain first motion information of an eyeball of a user when a shooting preview interface is displayed, where the shooting preview interface includes N preview sub-areas, each preview sub-area is associated with one camera, and N is a positive integer;
a
first processing module702, configured to determine, based on the first motion information, target preview sub-areas at which the eyes of the user gaze, where the N preview sub-areas include the target preview sub-area;
the photographing
module703 is configured to control a target camera associated with the target preview sub-region to photograph a target image when a duration that the eyeballs of the user watch the target preview sub-region is within a preset duration range.
Optionally, each of the N preview sub-areas includes an area identifier, and the area identifier has a first display mode and a second display mode;
under the condition that the area identification is in the first display mode, the area identification is used for indicating the eyeball of the user to watch the preview sub-area where the area identification is located;
and under the condition that the area identification is in the second display mode, the area identification is used for indicating that eyeballs of the user do not watch the preview sub-area where the area identification is located.
Optionally, the apparatus 700 further comprises:
and the second processing module is used for updating the area identification in the target preview subarea into the first display mode.
Optionally, the shooting preview interface further includes a position identifier, where the position identifier is used to indicate first motion information of the user's eyes, and the first motion information includes a real-time position where the user's eyes gaze;
the
first processing module702 comprises:
the first processing unit is used for determining the first preview sub-area as a target preview sub-area watched by eyeballs of a user under the condition that the display position of the position mark is located in the display range of the first preview sub-area in a target time period;
wherein the first preview sub-area is any one of the N preview sub-areas.
Optionally, the apparatus 700 further comprises:
the first receiving module is used for receiving first input of a user to M preview sub-areas in the N preview sub-areas;
the third processing module is used for responding to the first input, combining the M preview sub-areas into a second preview sub-area, performing image synthesis on preview images acquired by the M cameras displayed in the M preview sub-areas, generating a target preview image, and displaying the target preview image in the second preview sub-area;
wherein M is an integer greater than 1.
Optionally, the apparatus 700 further comprises:
the second receiving module is used for receiving a second input of the user to the second preview subarea;
and the fourth processing module is configured to split the second preview sub-area into M preview sub-areas in response to the second input, and display a preview image acquired by each of the M cameras in each of the M preview sub-areas.
Optionally, under the condition that a preview image displayed by a third preview sub-area of the N preview sub-areas is obtained by synthesizing preview images acquired by at least two cameras, the third preview sub-area further includes a camera number identifier, and the camera number identifier is used for indicating the number of cameras associated with the third preview sub-area.
Optionally, the apparatus 700 further comprises:
the fifth processing module is used for determining a camera obtained by recognizing the voice information as a target camera under the condition of acquiring the voice information input by the user; or,
the sixth processing module is used for determining a camera associated with a fourth preview sub-area as a target camera under the condition that selection input of a user on the fourth preview sub-area in the N preview sub-areas is received;
and the preset first time length is greater than or equal to the maximum value of the preset time length range.
Optionally, the apparatus 700 further comprises:
the first display module is used for displaying a shooting parameter setting window under the condition that third input of a user to the target preview sub-area is received, wherein the shooting parameter setting window comprises at least one shooting parameter;
the third receiving module is used for receiving fourth input of the shooting parameter setting window by a user;
a seventh processing module, configured to determine a first shooting parameter in response to the fourth input;
the photographing
module703 includes:
and the first photographing unit is used for controlling the target camera associated with the target preview sub-area to photograph the target image according to the first photographing parameter.
Optionally, the apparatus 700 further comprises:
the second acquisition module is used for acquiring second motion information of eyeballs of the user;
the eighth processing module controls the position identifier to move according to the second motion information;
the ninth processing module is used for determining a second shooting parameter according to the moving distance of the position identifier;
the photographing
module703 includes:
and the second photographing unit is used for controlling the target camera associated with the target preview sub-area to photograph the target image according to the second photographing parameter.
The photographing device in the embodiment of the present application may be a device, or may be a component, an integrated circuit, or a chip in a terminal. The device can be mobile electronic equipment or non-mobile electronic equipment. By way of example, the mobile electronic device may be a mobile phone, a tablet computer, a notebook computer, a palm top computer, a vehicle-mounted electronic device, a wearable device, an ultra-mobile personal computer (UMPC), a netbook or a Personal Digital Assistant (PDA), and the like, and the non-mobile electronic device may be a Network Attached Storage (NAS), a personal computer (personal computer, PC), a Television (TV), a teller machine, a self-service machine, and the like, and the embodiments of the present application are not limited in particular.
The photographing device in the embodiment of the present application may be a device having an operating system. The operating system may be an Android (Android) operating system, an ios operating system, or other possible operating systems, and embodiments of the present application are not limited specifically.
The photographing device provided in the embodiment of the present application can implement each process implemented by the method embodiments of fig. 1 to 6, and is not described here again to avoid repetition.
According to the photographing device, the first movement information of the eyeballs of the user is acquired through the first acquisition module under the condition that the photographing preview interface is displayed, wherein the photographing preview interface comprises N preview sub-areas, each preview sub-area is associated with one camera, and N is a positive integer; the first processing module determines target preview sub-areas watched by eyeballs of the user based on the first motion information, wherein the N preview sub-areas comprise the target preview sub-areas; the photographing module controls the target camera associated with the target preview subarea to photograph the target image under the condition that the time length of the target preview subarea watched by the eyeballs of the user is within the preset time length range, so that the camera required by the user can be conveniently and quickly selected to photograph through the movement of the eyeballs of the user, and the photographing mode is more flexible.
Optionally, as shown in fig. 8, an
electronic device800 is further provided in this embodiment of the present application, and includes a
processor801, a
memory802, and a program or an instruction stored in the
memory802 and executable on the
processor801, where the program or the instruction is executed by the
processor801 to implement each process of the foregoing photographing method embodiment, and can achieve the same technical effect, and in order to avoid repetition, details are not repeated here.
It should be noted that the electronic device in the embodiment of the present application includes the mobile electronic device and the non-mobile electronic device described above.
Fig. 9 is a schematic diagram of a hardware structure of an electronic device implementing various embodiments of the present application.
The
electronic device900 includes, but is not limited to: a
radio frequency unit901, a
network module902, an
audio output unit903, an
input unit904, a
sensor905, a
display unit906, a
user input unit907, an
interface unit908, a
memory909, a
processor910, and a
power supply911.
Those skilled in the art will appreciate that the
electronic device900 may further include a power source (e.g., a battery) for supplying power to various components, and the power source may be logically connected to the
processor910 through a power management system, so as to manage charging, discharging, and power consumption management functions through the power management system. The electronic device structure shown in fig. 9 does not constitute a limitation of the electronic device, and the electronic device may include more or less components than those shown, or combine some components, or arrange different components, and thus, the description is not repeated here.
The
processor910 is configured to obtain first motion information of an eyeball of a user under a condition that a shooting preview interface is displayed, where the shooting preview interface includes N preview sub-areas, each preview sub-area is associated with one camera, and N is a positive integer; determining a target preview subregion of the user's eye gaze based on the first motion information, the N preview subregions including the target preview subregion; and under the condition that the time length of the eyeballs of the user watching the target preview subarea is within a preset time length range, controlling a target camera associated with the target preview subarea to shoot a target image.
In the embodiment of the application, the camera required by the user can be conveniently and quickly selected to take pictures through the movement of the eyeballs of the user, so that the picture taking mode is more flexible.
Optionally, each of the N preview sub-areas includes an area identifier, and the area identifier has a first display mode and a second display mode;
under the condition that the area identification is in the first display mode, the area identification is used for indicating the eyeball of the user to watch the preview sub-area where the area identification is located;
and under the condition that the area identification is in the second display mode, the area identification is used for indicating that eyeballs of the user do not watch the preview sub-area where the area identification is located.
Optionally, the
processor910 is further configured to:
and updating the area identification in the target preview subregion to the first display mode.
Optionally, the shooting preview interface further includes a position identifier, where the position identifier is used to indicate first motion information of the user's eyes, and the first motion information includes a real-time position where the user's eyes gaze;
processor910, further configured to:
in a target time period, under the condition that the display position of the position mark is located in the display range of a first preview subregion, determining the first preview subregion as a target preview subregion watched by eyeballs of a user;
wherein the first preview sub-area is any one of the N preview sub-areas.
Optionally, the
user input unit907 is configured to receive a first input of the user to M preview sub-regions of the N preview sub-regions;
processor910, further configured to:
in response to the first input, combining the M preview sub-areas into a second preview sub-area, performing image synthesis on preview images acquired by the M cameras displayed in the M preview sub-areas, generating a target preview image, and displaying the target preview image on the second preview sub-area through a
display unit906; wherein M is an integer greater than 1.
Optionally, the
user input unit907 is further configured to receive a second input from the user to the second preview sub-area; in response to the second input, splitting the second preview sub-area into M preview sub-areas, and displaying a preview image acquired by each camera in the M cameras in each preview sub-area in the M preview sub-areas through the
display unit906.
Optionally, under the condition that a preview image displayed by a third preview sub-area of the N preview sub-areas is obtained by synthesizing preview images acquired by at least two cameras, the third preview sub-area further includes a camera number identifier, and the camera number identifier is used for indicating the number of cameras associated with the third preview sub-area.
Optionally, the
processor910 is further configured to:
under the condition of acquiring voice information input by a user, determining a camera obtained by recognizing the voice information as a target camera; or, under the condition that a selection input of a user for a fourth preview sub-area in the N preview sub-areas is received, determining a camera associated with the fourth preview sub-area as a target camera; and the preset first time length is greater than or equal to the maximum value of the preset time length range.
Optionally, the
display unit906 is further configured to display a shooting parameter setting window in a case that a third input of the user to the target preview sub-area is received, where the shooting parameter setting window includes at least one shooting parameter; a
user input unit907 further configured to receive a fourth input of the shooting parameter setting window from the user; a
processor910, further configured to determine a first shooting parameter in response to the fourth input; and controlling a target camera associated with the target preview subregion to shoot a target image according to the first shooting parameter.
Optionally, the
processor910 is further configured to obtain second motion information of the user's eyeball; controlling the position mark to move according to the second motion information; determining a second shooting parameter according to the moving distance of the position identifier; and controlling the target camera associated with the target preview subarea to shoot the target image according to the second shooting parameter.
In the embodiment of the application, the camera required by the user can be conveniently and quickly selected to take pictures through the movement of the eyeballs of the user, so that the picture taking mode is more flexible.
It should be understood that, in the embodiment of the present application, the
input Unit904 may include a Graphics Processing Unit (GPU) 9041 and a
microphone9042, and the
Graphics Processing Unit9041 processes image data of a still picture or a video obtained by an image capturing device (such as a camera) in a video capturing mode or an image capturing mode. The
display unit906 may include a
display panel9061, and the
display panel9061 may be configured in the form of a liquid crystal display, an organic light emitting diode, or the like. The
user input unit907 includes a
touch panel9071 and
other input devices9072. A
touch panel9071 also referred to as a touch screen. The
touch panel9071 may include two parts, a touch detection device and a touch controller.
Other input devices9072 may include, but are not limited to, a physical keyboard, function keys (e.g., volume control keys, switch keys, etc.), a trackball, a mouse, and a joystick, which are not described in detail herein.
Memory909 can be used to store software programs as well as various data including, but not limited to, application programs and operating systems. The
processor910 may integrate an application processor, which primarily handles operating systems, user interfaces, applications, etc., and a modem processor, which primarily handles wireless communications. It is to be appreciated that the modem processor described above may not be integrated into
processor910.
The embodiment of the present application further provides a readable storage medium, where a program or an instruction is stored on the readable storage medium, and when the program or the instruction is executed by a processor, the program or the instruction implements each process of the foregoing photographing method embodiment, and can achieve the same technical effect, and in order to avoid repetition, details are not repeated here.
The processor is the processor in the electronic device described in the above embodiment. The readable storage medium includes a computer readable storage medium, such as a Read-Only Memory (ROM), a Random Access Memory (RAM), a magnetic disk or an optical disk, and so on.
The embodiment of the present application further provides a chip, where the chip includes a processor and a communication interface, the communication interface is coupled to the processor, and the processor is configured to run a program or an instruction to implement each process of the foregoing photographing method embodiment, and can achieve the same technical effect, and the details are not repeated here to avoid repetition.
It should be understood that the chips mentioned in the embodiments of the present application may also be referred to as system-on-chip, system-on-chip or system-on-chip, etc.
It should be noted that, in this document, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising an … …" does not exclude the presence of other like elements in a process, method, article, or apparatus that comprises the element. Further, it should be noted that the scope of the methods and apparatus of the embodiments of the present application is not limited to performing the functions in the order illustrated or discussed, but may include performing the functions in a substantially simultaneous manner or in a reverse order based on the functions involved, e.g., the methods described may be performed in an order different than that described, and various steps may be added, omitted, or combined. In addition, features described with reference to certain examples may be combined in other examples.
Through the above description of the embodiments, those skilled in the art will clearly understand that the method of the above embodiments can be implemented by software plus a necessary general hardware platform, and certainly can also be implemented by hardware, but in many cases, the former is a better implementation manner. Based on such understanding, the technical solutions of the present application may be embodied in the form of a software product, which is stored in a storage medium (such as ROM/RAM, magnetic disk, optical disk) and includes instructions for enabling a terminal (such as a mobile phone, a computer, a server, an air conditioner, or a network device) to execute the method according to the embodiments of the present application.
While the present embodiments have been described with reference to the accompanying drawings, it is to be understood that the invention is not limited to the precise embodiments described above, which are meant to be illustrative and not restrictive, and that various changes may be made therein by those skilled in the art without departing from the spirit and scope of the invention as defined by the appended claims.