Method and system for invisible light communication using visible light camera

文档序号:328343 发布日期:2021-11-30 浏览:11次 中文

阅读说明:本技术 采用可见光摄像机的不可见光通信的方法和系统 (Method and system for invisible light communication using visible light camera ) 是由 罗俊 于 2019-04-15 设计创作,主要内容包括:在一个实施例中,一种通信方法,包括:由包括可见光摄像机模块的接收设备执行接收方法,可见光摄像机模块能够撷取由发送设备的不可见光信号发射器发出的不可见光。所述接收方法包括:广播摄像机帧率;由可见光摄像机模块、以摄像机帧率撷取多个帧,其中每个帧包括多个信号脉冲的一部分的一个相应的信号脉冲所生成的一个相应的区域,不可见光信号发射器以与摄像机帧率基本相同的信号脉冲速率,发出信号脉冲;检测每个帧中的相应的区域;以及标记每个帧中相应的区域,以获得相应于信号脉冲图案的编码图案,并对编码图案进行解码,以识别出数据序列。(In one embodiment, a method of communication, comprising: the receiving method is performed by a receiving apparatus including a visible light camera module capable of capturing invisible light emitted by an invisible light signal emitter of a transmitting apparatus. The receiving method comprises the following steps: broadcasting a camera frame rate; capturing, by the visible light camera module, a plurality of frames at a camera frame rate, wherein each frame includes a respective region generated by a respective signal pulse of a portion of the plurality of signal pulses, the invisible light signal emitter emitting the signal pulses at a signal pulse rate substantially the same as the camera frame rate; detecting a corresponding region in each frame; and marking the corresponding region in each frame to obtain a coded pattern corresponding to the signal pulse pattern and decoding the coded pattern to identify the data sequence.)

1. A method of communication, comprising:

performing a transmission method by a transmission apparatus including a first invisible light signal transmitter, wherein the transmission method includes:

loading an encoded signal pulse pattern having a data sequence;

scanning a camera frame rate broadcast by a receiving device; and

issuing, by the first invisible light signal emitter, a plurality of first signal pulses at a signal pulse rate substantially the same as the camera frame rate, wherein the signal pulse pattern is repeatedly transmitted in the plurality of first signal pulses.

2. The communication method of claim 1, further comprising:

performing a receiving method by a receiving apparatus comprising a visible light camera module capable of capturing invisible light emitted by the first invisible light signal emitter, wherein the receiving method comprises:

broadcasting the camera frame rate;

capturing, by the visible light camera module, a plurality of first frames at the camera frame rate, wherein each first frame comprises a respective first region generated by a respective signal pulse of a first portion of the plurality of first signal pulses;

detecting the respective first region in each first frame; and

the corresponding first region in each first frame is marked to obtain a first coding pattern corresponding to the signal pulse pattern, and the first coding pattern is decoded to identify the data sequence.

3. The communication method according to claim 2, wherein

The receiving method further comprises the following steps before the step of capturing the first frame at the camera frame rate:

acquiring the visibility of an overexposed second area in a second frame captured at the exposure time interval of a shutter of the visible light camera module; and

reducing the exposure time interval so that visibility of a third area in a third frame captured at the reduced exposure time interval is higher than visibility of the second area; and

capturing the first frame using the reduced exposure time interval.

4. A communication method according to claim 3, wherein the reduced exposure time interval is reduced such that visibility of the third area is highest.

5. The communication method according to claim 2, wherein the receiving method further comprises the following steps before the step of capturing the first frame at the camera frame rate:

adjusting the white balance of the visible camera module such that the color of the non-visible light emitted by the first non-visible light signal emitter is more distinguishable from the color of visible light, which is similar to the color of the non-visible light emitted by the first non-visible light signal emitter at neutral white balance.

6. The communication method according to claim 5, wherein the invisible light emitted by the first invisible light signal emitter is infrared light.

7. The communication method according to claim 6, wherein the white balance is adjusted to a color temperature of less than 3000K.

8. The communication method according to claim 5, wherein the step of detecting the respective first region in each first frame comprises:

binarizing each first frame using a threshold value to obtain a binarized frame, the threshold value being within a neighborhood of a set of pixel values of a pixel in a first one of a plurality of first regions, wherein the first one of the plurality of first regions corresponds to a bright signal pulse of the plurality of first signal pulses, and the set of pixel values includes color values;

extracting at least one corresponding first bright region in each binarized frame to obtain a position of each of the at least one corresponding first bright regions; and

tracking a location of each of the at least one respective first bright regions to detect each of at least one set of bright and dark regions corresponding to the binarized frame and being location-dependent, wherein the at least one set of bright and dark regions includes the first region.

9. The communication method according to claim 2, wherein the receiving method further comprises the steps of:

prior to the step of capturing the first frame at the camera frame rate,

capturing a plurality of second frames at the camera frame rate, wherein each second frame includes a respective second region generated by a respective signal pulse of a second portion of the plurality of first signal pulses;

detecting the respective second region in each second frame to obtain a location of the respective second region; and

in the step of capturing the first frame at the camera frame rate,

identifying the first invisible light signal emitter as an object on object tracking by the location of each of at least one of the plurality of second areas to automatically focus on the first invisible light signal emitter when the first invisible light signal emitter emits the first portion of the first signal pulse.

10. The communication method according to claim 9, wherein

Each second frame further includes at least one respective third region corresponding to at least one other transmitting device, wherein each of the at least one respective third region is generated by a respective one of a plurality of second signal pulses emitted by the second invisible light signal emitter of a respective other transmitting device of the at least one other transmitting device, and the plurality of second signal pulses correspond to the plurality of second frames;

the step of detecting the corresponding second region in each second frame to obtain the location of the corresponding second region comprises:

detecting the respective second region and the at least one respective third region in each second frame to obtain a location of the respective second region and a location of each of the at least one respective third region; and

selecting a location of each of at least one of the plurality of second regions for autofocusing for the second region in a second frame of the plurality of second frames and a largest one of the at least one third region based on a region size of the second region in the second frame of the plurality of second frames.

11. The communication method according to claim 10, wherein the step of detecting the corresponding second area in each second frame to obtain the location of the corresponding second area further comprises:

calculating the number of the second region and the at least one third region in the second frame of the plurality of second frames to obtain the number of the transmitting device and the at least one other transmitting device; and

averaging, for each of the transmitting device and the at least one other transmitting device, a region size of a plurality of respective regions in the plurality of second frames based on the number of the transmitting device and the at least one other transmitting device, wherein each respective region in the plurality of respective regions is one of the second region and the at least one third region in a respective third frame of the plurality of second frames, the one of the second region and the at least one third region being generated by a bright signal pulse in the plurality of first signal pulses or by a bright signal pulse in the plurality of second signal pulses corresponding to one of the at least one third region.

12. The communication method according to claim 10, wherein

Each second frame further includes a respective fourth region corresponding to a portion of the transmitting device that does not include the first invisible light signal emitter, or a portion of any of the at least one other transmitting device that does not include the second invisible light signal emitter; and

aligning the respective second regions in each second frame with each other by aligning the respective fourth regions in each second frame with each other before the step of detecting the respective second regions in each second frame to obtain the positions of the respective second regions.

13. The communication method according to claim 9, wherein

Each second frame further includes a respective fourth region corresponding to a portion of the transmitting device that does not include the first invisible light signal emitter; and

aligning the respective second regions in each second frame with each other by aligning the respective fourth regions in each second frame with each other before the step of detecting the respective second regions in each second frame to obtain the positions of the respective second regions.

14. The communication method according to claim 2, wherein the step of detecting the respective first region in each first frame comprises:

binarizing each first frame using a threshold value to obtain a binarized frame, the threshold value being within a vicinity of at least one pixel value of one pixel in one of a plurality of first regions corresponding to one bright signal pulse among the plurality of first signal pulses;

extracting at least one corresponding first bright region in each binarized frame to obtain a position of each of the at least one corresponding first bright regions; and

tracking a location of each of the at least one respective first bright regions to detect each of at least one set of bright and dark regions corresponding to the binarized frame and being location-dependent, wherein the at least one set of bright and dark regions includes the first region.

15. The communication method of claim 14, wherein the at least one pixel value is a luminance value.

16. The communication method of claim 14, wherein said marking the corresponding first region in each first frame to obtain a first coding pattern corresponding to the signal pulse pattern and decoding the first coding pattern to identify the data sequence comprises:

marking the at least one set of light and dark regions to obtain a respective second coding pattern for each of the at least one set of light and dark regions; and

determining a third encoding pattern of the respective second encoding pattern comprising the first encoding pattern that repeats each of the at least one set of light and dark regions.

17. The communication method of claim 2, wherein when the first invisible light signal emitter emits the plurality of first signal pulses, the first invisible light signal emitter is not limited to emitting first signal pulses synchronized with the timing at which the visible light camera module captures the first frame.

18. A method of communication, comprising:

performing a receiving method by a receiving apparatus including a visible light camera module capable of capturing invisible light emitted by a first invisible light signal emitter of a transmitting apparatus, wherein the receiving method includes:

broadcasting a camera frame rate;

capturing, by the visible light camera module, a plurality of first frames at the camera frame rate, wherein each first frame includes a respective first region generated by a respective signal pulse of a first portion of the plurality of first signal pulses, the first invisible light signal emitter emitting the first signal pulses at a signal pulse rate substantially the same as the camera frame rate;

detecting the respective first region in each first frame; and

the corresponding first region in each first frame is marked to obtain a first coding pattern corresponding to the signal pulse pattern, and the first coding pattern is decoded to identify the data sequence.

19. The communication method of claim 18, further comprising:

performing a transmission method by the transmission apparatus including the first invisible light signal transmitter, wherein the transmission method includes:

loading the encoded signal pulse pattern with the data sequence;

scanning a camera frame rate broadcast by the receiving device; and

emitting, by the first invisible light signal emitter, the first plurality of signal pulses at the signal pulse rate, wherein the signal pulse pattern is repeatedly transmitted in the first plurality of signal pulses.

20. The communication method of claim 19, wherein

The receiving method further comprises the following steps before the step of capturing the first frame at the camera frame rate:

acquiring the visibility of an overexposed second area in a second frame captured at the exposure time interval of a shutter of the visible light camera module; and

reducing the exposure time interval so that visibility of a third area in a third frame captured at the reduced exposure time interval is higher than visibility of the second area; and

capturing the first frame using the reduced exposure time interval.

21. The communication method according to claim 20, wherein the reduced exposure time interval is reduced such that visibility of the third area is highest.

22. The communication method of claim 19, wherein the receiving method further comprises the following steps before the step of capturing the first frame at the camera frame rate:

adjusting the white balance of the visible camera module such that the color of the non-visible light emitted by the first non-visible light signal emitter is more distinguishable from the color of visible light, which is similar to the color of the non-visible light emitted by the first non-visible light signal emitter at neutral white balance.

23. The communication method according to claim 22, wherein the invisible light emitted by the first invisible light signal emitter is infrared light.

24. The communication method of claim 23, wherein the white balance is adjusted to a color temperature of less than 3000K.

25. The communication method of claim 22, wherein the step of detecting the respective first region in each first frame comprises:

binarizing each first frame using a threshold value to obtain a binarized frame, the threshold value being within a neighborhood of a set of pixel values of a pixel in a first one of a plurality of first regions, wherein the first one of the plurality of first regions corresponds to a bright signal pulse of the plurality of first signal pulses, and the set of pixel values includes color values;

extracting at least one corresponding first bright region in each binarized frame to obtain a position of each of the at least one corresponding first bright regions; and

tracking a location of each of the at least one respective first bright regions to detect each of at least one set of bright and dark regions corresponding to the binarized frame and being location-dependent, wherein the at least one set of bright and dark regions includes the first region.

26. The communication method of claim 19, wherein the receiving method further comprises the steps of:

prior to the step of capturing the first frame at the camera frame rate,

capturing a plurality of second frames at the camera frame rate, wherein each second frame includes a respective second region generated by a respective signal pulse of a second portion of the plurality of first signal pulses;

detecting the respective second region in each second frame to obtain a location of the respective second region; and

in the step of capturing the first frame at the camera frame rate,

identifying the first invisible light signal emitter as an object on object tracking by the location of each of at least one of the plurality of second areas to automatically focus on the first invisible light signal emitter when the first invisible light signal emitter emits the first portion of the first signal pulse.

27. The communication method of claim 26, wherein

Each second frame further includes at least one respective third region corresponding to at least one other transmitting device, wherein each of the at least one respective third region is generated by a respective one of a plurality of second signal pulses emitted by the second invisible light signal emitter of a respective other transmitting device of the at least one other transmitting device, and the plurality of second signal pulses correspond to the plurality of second frames;

the step of detecting the corresponding second region in each second frame to obtain the location of the corresponding second region comprises:

detecting the respective second region and the at least one respective third region in each second frame to obtain a location of the respective second region and a location of each of the at least one respective third region; and

selecting a location of each of at least one of the plurality of second regions for autofocusing for the second region in a second frame of the plurality of second frames and a largest one of the at least one third region based on a region size of the second region in the second frame of the plurality of second frames.

28. The communication method according to claim 27, wherein the step of detecting the corresponding second area in each second frame to obtain the location of the corresponding second area further comprises:

calculating the number of the second region and the at least one third region in the second frame of the plurality of second frames to obtain the number of the transmitting device and the at least one other transmitting device; and

averaging, for each of the transmitting device and the at least one other transmitting device, a region size of a plurality of respective regions in the plurality of second frames based on the number of the transmitting device and the at least one other transmitting device, wherein each respective region in the plurality of respective regions is one of the second region and the at least one third region in a respective third frame of the plurality of second frames, the one of the second region and the at least one third region being generated by a bright signal pulse in the plurality of first signal pulses or by a bright signal pulse in the plurality of second signal pulses corresponding to one of the at least one third region.

29. The communication method of claim 27, wherein

Each second frame further includes a respective fourth region corresponding to a portion of the transmitting device that does not include the first invisible light signal emitter, or a portion of any of the at least one other transmitting device that does not include the second invisible light signal emitter; and

aligning the respective second regions in each second frame with each other by aligning the respective fourth regions in each second frame with each other before the step of detecting the respective second regions in each second frame to obtain the positions of the respective second regions.

30. The communication method of claim 26, wherein

Each second frame further includes a respective fourth region corresponding to a portion of the transmitting device that does not include the first invisible light signal emitter; and

aligning the respective second regions in each second frame with each other by aligning the respective fourth regions in each second frame with each other before the step of detecting the respective second regions in each second frame to obtain the positions of the respective second regions.

31. The communication method of claim 19, wherein the step of detecting the respective first region in each first frame comprises:

binarizing each first frame using a threshold value to obtain a binarized frame, the threshold value being within a vicinity of at least one pixel value of one pixel in one of a plurality of first regions corresponding to one bright signal pulse among the plurality of first signal pulses;

extracting at least one corresponding first bright region in each binarized frame to obtain a position of each of the at least one corresponding first bright regions; and

tracking a location of each of the at least one respective first bright regions to detect each of at least one set of bright and dark regions corresponding to the binarized frame and being location-dependent, wherein the at least one set of bright and dark regions includes the first region.

32. The communication method of claim 31, wherein the at least one pixel value is a luminance value.

33. The communication method of claim 31, wherein said marking the corresponding first region in each first frame to obtain a first coding pattern corresponding to the signal pulse pattern and decoding the first coding pattern to identify the data sequence comprises:

marking the at least one set of light and dark regions to obtain a respective second coding pattern for each of the at least one set of light and dark regions; and

determining a third encoding pattern of the respective second encoding pattern comprising the first encoding pattern that repeats each of the at least one set of light and dark regions.

34. The communication method of claim 19, wherein when the first invisible light signal emitter emits the plurality of first signal pulses, the first invisible light signal emitter is not limited to emitting first signal pulses synchronized with the timing at which the visible light camera module captures the first frame.

35. A communication system, comprising:

a transmitting device, comprising:

a first invisible light signal emitter;

at least a first memory configured to store first program instructions; and

at least a first processor configured to execute the first program instructions that cause a transmission method to be performed, wherein the transmission method comprises:

loading an encoded signal pulse pattern having a data sequence;

scanning a camera frame rate broadcast by a receiving device; and

issuing, by the first invisible light signal emitter, a plurality of first signal pulses at a signal pulse rate substantially the same as the camera frame rate, wherein the signal pulse pattern is repeatedly transmitted in the plurality of first signal pulses.

36. The communication system of claim 35, further comprising:

a receiving apparatus, comprising:

a visible light camera module capable of capturing invisible light emitted by the first invisible light signal emitter;

at least one second memory configured to store second program instructions; and

at least a second processor configured to execute the second program instructions that cause a receiving method to be performed, wherein the receiving method comprises:

broadcasting the camera frame rate;

capturing, by the visible light camera module, a plurality of first frames at the camera frame rate, wherein each first frame comprises a respective first region generated by a respective signal pulse of a first portion of the plurality of first signal pulses;

detecting the respective first region in each first frame; and

the corresponding first region in each first frame is marked to obtain a first coding pattern corresponding to the signal pulse pattern, and the first coding pattern is decoded to identify the data sequence.

37. The communication system of claim 36, wherein

The receiving method further comprises the following steps before the step of capturing the first frame at the camera frame rate:

acquiring the visibility of an overexposed second area in a second frame captured at the exposure time interval of a shutter of the visible light camera module; and

reducing the exposure time interval so that visibility of a third area in a third frame captured at the reduced exposure time interval is higher than visibility of the second area; and

capturing the first frame using the reduced exposure time interval.

38. The communication system of claim 37, wherein the reduced exposure time interval is reduced such that visibility of the third area is highest.

39. The communication system of claim 36, wherein the receiving method further comprises, before the step of capturing the first frame at the camera frame rate:

adjusting the white balance of the visible camera module such that the color of the non-visible light emitted by the first non-visible light signal emitter is more distinguishable from the color of visible light, which is similar to the color of the non-visible light emitted by the first non-visible light signal emitter at neutral white balance.

40. The communication system of claim 39, wherein the first invisible light signal emitter emits invisible light that is infrared.

41. The communication system of claim 40, wherein the white balance is adjusted to a color temperature of less than 3000K.

42. The communication system of claim 39, wherein the step of detecting the respective first region in each first frame comprises:

binarizing each first frame using a threshold value to obtain a binarized frame, the threshold value being within a neighborhood of a set of pixel values of a pixel in a first one of a plurality of first regions, wherein the first one of the plurality of first regions corresponds to a bright signal pulse of the plurality of first signal pulses, and the set of pixel values includes color values;

extracting at least one corresponding first bright region in each binarized frame to obtain a position of each of the at least one corresponding first bright regions; and

tracking a location of each of the at least one respective first bright regions to detect each of at least one set of bright and dark regions corresponding to the binarized frame and being location-dependent, wherein the at least one set of bright and dark regions includes the first region.

43. The communication system of claim 36, wherein the receiving method further comprises the steps of:

prior to the step of capturing the first frame at the camera frame rate,

capturing a plurality of second frames at the camera frame rate, wherein each second frame includes a respective second region generated by a respective signal pulse of a second portion of the plurality of first signal pulses;

detecting the respective second region in each second frame to obtain a location of the respective second region; and

in the step of capturing the first frame at the camera frame rate,

identifying the first invisible light signal emitter as an object on object tracking by the location of each of at least one of the plurality of second areas to automatically focus on the first invisible light signal emitter when the first invisible light signal emitter emits the first portion of the first signal pulse.

44. The communication system of claim 43, wherein

Each second frame further includes at least one respective third region corresponding to at least one other transmitting device, wherein each of the at least one respective third region is generated by a respective one of a plurality of second signal pulses emitted by the second invisible light signal emitter of a respective other transmitting device of the at least one other transmitting device, and the plurality of second signal pulses correspond to the plurality of second frames;

the step of detecting the corresponding second region in each second frame to obtain the location of the corresponding second region comprises:

detecting the respective second region and the at least one respective third region in each second frame to obtain a location of the respective second region and a location of each of the at least one respective third region; and

selecting a location of each of at least one of the plurality of second regions for autofocusing for the second region in a second frame of the plurality of second frames and a largest one of the at least one third region based on a region size of the second region in the second frame of the plurality of second frames.

45. The communication system of claim 44, wherein the step of detecting the corresponding second region in each second frame to obtain the location of the corresponding second region further comprises:

calculating the number of the second region and the at least one third region in the second frame of the plurality of second frames to obtain the number of the transmitting device and the at least one other transmitting device; and

averaging, for each of the transmitting device and the at least one other transmitting device, a region size of a plurality of respective regions in the plurality of second frames based on the number of the transmitting device and the at least one other transmitting device, wherein each respective region in the plurality of respective regions is one of the second region and the at least one third region in a respective third frame of the plurality of second frames, the one of the second region and the at least one third region being generated by a bright signal pulse in the plurality of first signal pulses or by a bright signal pulse in the plurality of second signal pulses corresponding to one of the at least one third region.

46. The communication system of claim 44, wherein

Each second frame further includes a respective fourth region corresponding to a portion of the transmitting device that does not include the first invisible light signal emitter, or a portion of any of the at least one other transmitting device that does not include the second invisible light signal emitter; and

aligning the respective second regions in each second frame with each other by aligning the respective fourth regions in each second frame with each other before the step of detecting the respective second regions in each second frame to obtain the positions of the respective second regions.

47. The communication system of claim 43, wherein

Each second frame further includes a respective fourth region corresponding to a portion of the transmitting device that does not include the first invisible light signal emitter; and

aligning the respective second regions in each second frame with each other by aligning the respective fourth regions in each second frame with each other before the step of detecting the respective second regions in each second frame to obtain the positions of the respective second regions.

48. The communication system of claim 36, wherein the step of detecting the respective first region in each first frame comprises:

binarizing each first frame using a threshold value to obtain a binarized frame, the threshold value being within a vicinity of at least one pixel value of one pixel in one of a plurality of first regions corresponding to one bright signal pulse among the plurality of first signal pulses;

extracting at least one corresponding first bright region in each binarized frame to obtain a position of each of the at least one corresponding first bright regions; and

tracking a location of each of the at least one respective first bright regions to detect each of at least one set of bright and dark regions corresponding to the binarized frame and being location-dependent, wherein the at least one set of bright and dark regions includes the first region.

49. The communication system of claim 48, wherein the at least one pixel value is a luminance value.

50. The communication system of claim 48, wherein said marking the corresponding first region in each first frame to obtain a first coding pattern corresponding to the signal pulse pattern and decoding the first coding pattern to identify the data sequence comprises:

marking the at least one set of light and dark regions to obtain a respective second coding pattern for each of the at least one set of light and dark regions; and

determining a third encoding pattern of the respective second encoding pattern comprising the first encoding pattern that repeats each of the at least one set of light and dark regions.

51. The communication system of claim 36, wherein when the first invisible light signal emitter emits the plurality of first signal pulses, the first invisible light signal emitter is not limited to emitting first signal pulses synchronized with the timing of the visible light camera module's acquisition of the first frame.

52. A communication system, comprising:

a receiving apparatus, comprising:

the visible light camera module can capture invisible light emitted by the first invisible light signal emitter of the sending equipment;

at least a first memory configured to store first program instructions; and

at least a first processor configured to execute the first program instructions that cause a receiving method to be performed, wherein the receiving method comprises:

broadcasting a camera frame rate;

capturing, by the visible light camera module, a plurality of first frames at the camera frame rate, wherein each first frame includes a respective first region generated by a respective signal pulse of a first portion of the plurality of first signal pulses, the first invisible light signal emitter emitting the first signal pulses at a signal pulse rate substantially the same as the camera frame rate;

detecting the respective first region in each first frame; and

the corresponding first region in each first frame is marked to obtain a first coding pattern corresponding to the signal pulse pattern, and the first coding pattern is decoded to identify the data sequence.

53. The communication system of claim 52, further comprising:

a transmitting device, comprising:

a first invisible light signal emitter;

at least one second memory configured to store second program instructions; and

at least a second processor configured to execute the second program instructions that cause a transmission method to be performed, wherein the transmission method comprises:

loading the encoded signal pulse pattern with the data sequence;

scanning a camera frame rate broadcast by the receiving device; and

emitting, by the first invisible light signal emitter, the first plurality of signal pulses at the signal pulse rate, wherein the signal pulse pattern is repeatedly transmitted in the first plurality of signal pulses.

54. A communication system according to claim 53, wherein the reduced exposure time interval is reduced such that visibility of the third area is highest.

55. The communication system of claim 52, wherein the receiving method further comprises, before the step of capturing the first frame at the camera frame rate:

adjusting the white balance of the visible camera module such that the color of the non-visible light emitted by the first non-visible light signal emitter is more distinguishable from the color of visible light, which is similar to the color of the non-visible light emitted by the first non-visible light signal emitter at neutral white balance.

56. The communication system of claim 55, wherein the first invisible light signal emitter emits invisible light that is infrared.

57. The communication system of claim 56, wherein the white balance is adjusted to a color temperature of less than 3000K.

58. The communication system of claim 55, wherein the step of detecting the respective first region in each first frame comprises:

binarizing each first frame using a threshold value to obtain a binarized frame, the threshold value being within a neighborhood of a set of pixel values of a pixel in a first one of a plurality of first regions, wherein the first one of the plurality of first regions corresponds to a bright signal pulse of the plurality of first signal pulses, and the set of pixel values includes color values;

extracting at least one corresponding first bright region in each binarized frame to obtain a position of each of the at least one corresponding first bright regions; and

tracking a location of each of the at least one respective first bright regions to detect each of at least one set of bright and dark regions corresponding to the binarized frame and being location-dependent, wherein the at least one set of bright and dark regions includes the first region.

59. The communication system of claim 52, wherein the receiving method further comprises the steps of:

prior to the step of capturing the first frame at the camera frame rate,

capturing a plurality of second frames at the camera frame rate, wherein each second frame includes a respective second region generated by a respective signal pulse of a second portion of the plurality of first signal pulses;

detecting the respective second region in each second frame to obtain a location of the respective second region; and

in the step of capturing the first frame at the camera frame rate,

identifying the first invisible light signal emitter as an object on object tracking by the location of each of at least one of the plurality of second areas to automatically focus on the first invisible light signal emitter when the first invisible light signal emitter emits the first portion of the first signal pulse.

60. The communication system of claim 59, wherein

Each second frame further includes at least one respective third region corresponding to at least one other transmitting device, wherein each of the at least one respective third region is generated by a respective one of a plurality of second signal pulses emitted by the second invisible light signal emitter of a respective other transmitting device of the at least one other transmitting device, and the plurality of second signal pulses correspond to the plurality of second frames;

the step of detecting the corresponding second region in each second frame to obtain the location of the corresponding second region comprises:

detecting the respective second region and the at least one respective third region in each second frame to obtain a location of the respective second region and a location of each of the at least one respective third region; and

selecting a location of each of at least one of the plurality of second regions for autofocusing for the second region in a second frame of the plurality of second frames and a largest one of the at least one third region based on a region size of the second region in the second frame of the plurality of second frames.

61. The communication system of claim 60, wherein the step of detecting the corresponding second region in each second frame to obtain the location of the corresponding second region further comprises:

calculating the number of the second region and the at least one third region in the second frame of the plurality of second frames to obtain the number of the transmitting device and the at least one other transmitting device; and

averaging, for each of the transmitting device and the at least one other transmitting device, a region size of a plurality of respective regions in the plurality of second frames based on the number of the transmitting device and the at least one other transmitting device, wherein each respective region in the plurality of respective regions is one of the second region and the at least one third region in a respective third frame of the plurality of second frames, the one of the second region and the at least one third region being generated by a bright signal pulse in the plurality of first signal pulses or by a bright signal pulse in the plurality of second signal pulses corresponding to one of the at least one third region.

62. The communication system of claim 60, wherein

Each second frame further includes a respective fourth region corresponding to a portion of the transmitting device that does not include the first invisible light signal emitter, or a portion of any of the at least one other transmitting device that does not include the second invisible light signal emitter; and

aligning the respective second regions in each second frame with each other by aligning the respective fourth regions in each second frame with each other before the step of detecting the respective second regions in each second frame to obtain the positions of the respective second regions.

63. The communication system of claim 59, wherein

Each second frame further includes a respective fourth region corresponding to a portion of the transmitting device that does not include the first invisible light signal emitter; and

aligning the respective second regions in each second frame with each other by aligning the respective fourth regions in each second frame with each other before the step of detecting the respective second regions in each second frame to obtain the positions of the respective second regions.

64. The communication system of claim 52, wherein the step of detecting the respective first region in each first frame comprises:

binarizing each first frame using a threshold value to obtain a binarized frame, the threshold value being within a vicinity of at least one pixel value of one pixel in one of a plurality of first regions corresponding to one bright signal pulse among the plurality of first signal pulses;

extracting at least one corresponding first bright region in each binarized frame to obtain a position of each of the at least one corresponding first bright regions; and

tracking a location of each of the at least one respective first bright regions to detect each of at least one set of bright and dark regions corresponding to the binarized frame and being location-dependent, wherein the at least one set of bright and dark regions includes the first region.

65. The communication system of claim 64, wherein the at least one pixel value is a luminance value.

66. The communication system of claim 64, wherein the step of marking the respective first region in each first frame to obtain a first coding pattern corresponding to the signal pulse pattern and decoding the first coding pattern to identify the data sequence comprises:

marking the at least one set of light and dark regions to obtain a respective second coding pattern for each of the at least one set of light and dark regions; and

determining a third encoding pattern of the respective second encoding pattern comprising the first encoding pattern that repeats each of the at least one set of light and dark regions.

67. The communication system of claim 52, wherein when the first invisible light signal emitter emits the plurality of first signal pulses, the first invisible light signal emitter is not limited to emitting first signal pulses synchronized with the timing of the visible light camera module capturing the first frame.

1. Field of the invention

The present application relates to the field of invisible light communication, and more particularly, to a method and system for invisible light communication using a visible light camera.

2. Description of the related Art

For visible or invisible light communication, the communication system comprises a transmitting device comprising a visible or invisible light signal emitter which transmits encoded light pulses with a data sequence, a receiving device comprising a camera which acquires frames correspondingly reflecting the sampled light pulses, and at least one processor which processes the frames to obtain the data sequence. However, in the case of using a visible light signal transmitter, the user's eyes may feel uncomfortable when seeing the light pulse. In the case of an invisible light signal transmitter, the camera is typically an invisible light camera, which may thus add additional cost to the receiving device.

Disclosure of Invention

It is an object of the present application to propose a method and system for invisible light communication using a visible light camera.

In a first aspect of the present application, a communication method includes: the transmission method is performed by a transmission apparatus including a first invisible light signal emitter. The sending method comprises the following steps: loading an encoded signal pulse pattern having a data sequence; scanning a camera frame rate broadcast by a receiving device; and issuing, by the first invisible light signal emitter, a plurality of first signal pulses at a signal pulse rate substantially the same as the camera frame rate, wherein the signal pulse pattern is repeatedly transmitted in the plurality of first signal pulses.

In a second aspect of the present application, a communication method includes: the receiving method is performed by a receiving apparatus comprising a visible light camera module capable of capturing invisible light emitted by a first invisible light signal emitter of a transmitting apparatus. The receiving method comprises the following steps: broadcasting a camera frame rate; capturing, by the visible light camera module, a plurality of first frames at the camera frame rate, wherein each first frame includes a respective first region generated by a respective signal pulse of a first portion of the plurality of first signal pulses, the first invisible light signal emitter emitting the first signal pulses at a signal pulse rate substantially the same as the camera frame rate; detecting the respective first region in each first frame; and marking the corresponding first region in each first frame to obtain a first coding pattern corresponding to the signal pulse pattern and decoding the first coding pattern to identify the data sequence.

In a third aspect of the present application, a communication system includes: and a sending device. The transmission apparatus includes: the device comprises a first invisible light signal emitter, at least one first memory and at least one first processor. The at least one first memory is configured to store first program instructions. The at least one first processor is configured to execute the first program instructions, which cause a transmission method to be performed. The sending method comprises the following steps: loading an encoded signal pulse pattern having a data sequence; scanning a camera frame rate broadcast by a receiving device; and issuing, by the first invisible light signal emitter, a plurality of first signal pulses at a signal pulse rate substantially the same as the camera frame rate, wherein the signal pulse pattern is repeatedly transmitted in the plurality of first signal pulses.

In a fourth aspect of the present application, a communication system includes: and receiving the device. The receiving apparatus includes: the device comprises a visible light camera module, at least one first memory and at least one first processor. The visible light camera module can capture invisible light emitted by the first invisible light signal emitter of the sending device. The at least one first memory is configured to store first program instructions. The at least one first processor is configured to execute the first program instructions, which cause the receiving method to be performed. The receiving method comprises the following steps: broadcasting a camera frame rate; capturing, by the visible light camera module, a plurality of first frames at the camera frame rate, wherein each first frame includes a respective first region generated by a respective signal pulse of a first portion of the plurality of first signal pulses, the first invisible light signal emitter emitting the first signal pulses at a signal pulse rate substantially the same as the camera frame rate; detecting the respective first region in each first frame; and marking the corresponding first region in each first frame to obtain a first coding pattern corresponding to the signal pulse pattern and decoding the first coding pattern to identify the data sequence.

Brief description of the drawings

In order to more clearly illustrate the embodiments of this application or the related art, the following drawings will be described in the embodiments and briefly introduced as follows. It should be apparent that the drawings represent only some of the embodiments of the present application and that other drawings may be derived by those skilled in the art from these drawings without making any preconditions.

Fig. 1 shows a block diagram of input, control and output hardware modules in a transmitting device according to an embodiment of the application.

Fig. 2 shows a block diagram of input, control and output hardware modules in a receiving device according to an embodiment of the application.

Fig. 3 shows a flowchart of a communication method of a transmitting device and a receiving device according to an embodiment of the present application.

Fig. 4 shows a schematic diagram of one of a plurality of first frames captured by a visible light camera module of a receiving apparatus according to an embodiment of the present application.

Fig. 5 shows a timing diagram of a synchronization start timing of a first frame at a signal pulse rate substantially the same as the camera frame rate for a first portion of a first signal pulse and for which sampling is successful, according to an embodiment of the application.

Fig. 6 shows a timing diagram of the signal pulse rate substantially the same as the camera frame rate, the start timing for the first portion of the first signal pulse and the first frame to make the sampling successful in a first asynchronous manner, according to an embodiment of the present application.

Fig. 7 shows a timing diagram of the start timing of a first frame at a signal pulse rate substantially the same as the camera frame rate for a first portion of a first signal pulse and for sampling to be successful in a second asynchronous manner, according to an embodiment of the application.

Fig. 8 shows a timing diagram of the signal pulse rate substantially the same as the camera frame rate, for a first portion of a first signal pulse and the start timing of a first frame in a third asynchronous manner to make the sampling successful according to an embodiment of the application.

FIG. 9 shows a timing diagram of the start timing of a first frame at a signal pulse rate substantially the same as the camera frame rate, for a first portion of a first signal pulse, and for which sampling may be unsuccessful in a fourth asynchronous manner, according to an embodiment of the application.

Fig. 10 shows a flow chart of an exposure time interval reduction method according to an embodiment of the present application to improve the unsuccessful sampling condition of fig. 9 and to improve the visibility of the portion of the first frame corresponding to the first region of the bright signal pulses of the first portion of the first signal pulses.

Fig. 11 shows a schematic diagram of an exemplary electro-optical characteristic curve of a visible-light camera for different exposure time intervals according to an embodiment of the application, wherein a reduced exposure time interval improves the visibility of a first region of a first frame of bright signal pulses corresponding to a first portion of a first signal pulse.

FIG. 12 shows a timing diagram of a reduced exposure time interval to improve the sampling failure condition of FIG. 9, according to an embodiment of the present application.

Fig. 13 is a flowchart illustrating a color conversion method according to an embodiment of the present application, in which a portion of a first region is more distinguishable from other regions of a first frame, and if the color conversion method 1300 is not performed, the other portions of the first frame have colors similar to the portion of the first region.

FIG. 14 shows a schematic diagram of an exemplary color conversion scene according to an embodiment of the present application.

Fig. 15 shows a flow chart of an auto-focusing method for auto-focusing on a first invisible light signal emitter when the first invisible light signal emitter emits a first portion of a first signal pulse according to an embodiment of the present application.

FIG. 16 is a diagram illustrating the steps of capturing, aligning, and detecting in the auto-focusing method shown in FIG. 15 according to an embodiment of the present disclosure.

Fig. 17 is a schematic diagram illustrating an auto-focusing step in the auto-focusing method shown in fig. 15 according to an embodiment of the present application.

Fig. 18 shows a flow chart of a method of autofocusing in the case of multiple transmitting devices according to an embodiment of the application.

FIG. 19 shows a schematic diagram of an exemplary multi-send device scenario to which the auto-focusing method of FIG. 18 may be applied, according to an embodiment of the present application.

Fig. 20 shows a flow chart of the detection step in the method of analyzing the first frame to obtain the data sequence according to an embodiment of the present application.

Fig. 21 shows a schematic diagram of a binarization step of removing a region other than the first region based on brightness or color in the detection step according to an embodiment of the present application.

Fig. 22 shows a schematic diagram of an extraction step and a tracking step of removing a region other than the first region based on the flicker characteristic in the detection step according to the embodiment of the present application.

Fig. 23 shows a flow chart of the marking and decoding steps in a method of analyzing a first frame to obtain a data sequence according to an embodiment of the application.

Fig. 24 shows a schematic diagram of a marking step and a determination step of removing a region that is not a first region based on a repetition characteristic and a decoding step in the marking and decoding step according to an embodiment of the present application.

Detailed Description

Embodiments of the present application will be described in detail below with reference to the accompanying drawings, wherein the embodiments are described in detail with reference to the accompanying drawings. In particular, the terminology used in the embodiments of the present application is for the purpose of describing particular embodiments only and is not intended to be limiting of the invention.

As used herein, the term "use" refers to the case where an object is directly employed to perform a step, or to the case where the object is modified through at least one intermediate step and the modified object is directly employed to perform a step.

Fig. 1 shows a block diagram of input, control and output hardware modules in a transmitting device 100 according to an embodiment of the application. Referring to fig. 1, the transmitting apparatus 100 includes a storage module 102, a wired or wireless communication module 104, a short-range wireless communication module 106, an invisible light signal transmitter 108, a processor module 110, a memory module 112, and a bus 114. The transmitting device 100 may be an internet-of-things (IoT) device, such as an intelligent home device, or an intelligent vehicle-mounted device. For example, the smart home device may be a smart television, a smart speaker, or a smart printer. For example, the smart in-vehicle device may be a Global Positioning System (GPS) receiver or an in-vehicle infotainment device.

The memory module 102 is an input hardware module and is configured to store a signal pulse pattern to be loaded into the transmission method 300 to be described with reference to fig. 3. The transmission method 300 is performed by the transmission apparatus 100. The signal pulse pattern is transmitted to the processor module 110 in a loading manner. Alternatively, another input hardware module may be used to input the signal pulse pattern, such as wired or wireless communication module 104. The wired or wireless communication module 104 is used to receive signal pulse patterns from the network through wired or wireless communication and transmit the signal pulse patterns to the processor module 110 by means of loading.

The short-range wireless communication module 106 may serve as an input hardware module and an output hardware module. The short range wireless communication module 106 performs the step of scanning for an advertisement message, the step of receiving a pairing request, the step of transmitting a pairing request response, and the step of pairing with a receiving device in the transmission method 300 using a short range wireless communication technology such as Bluetooth Low Energy (BLE). For example, the advertisement message is to be sent to the processor module 110. Upon receiving the instruction from processor module 110, short-range wireless communication module 106 uses BLE for communication. Alternatively, another short-range wireless communication technology, such as ZigBee, may be used.

The invisible light signal emitter 108 is an output hardware module. The invisible light signal emitter 108 is used in the transmission method 300 to transmit a first signal pulse. Upon receiving an instruction from the processor module 110, the invisible light signal emitter 108 emits the first signal pulse. The invisible light emitted by the invisible light signal emitter 108 may be infrared light. Alternatively, the invisible light emitted by the invisible light signal emitter 108 may be ultraviolet light.

Memory module 112 may be a transitory or non-transitory computer readable medium that includes at least one flash memory storing program instructions that, when executed by processor module 110, cause processor module 110 to control storage module 102, invisible light signal transmitter 108, and short-range wireless communication module 106 to perform transmission method 300. Processor module 110 includes at least one processor that sends signals to and/or receives signals from storage module 102, wired or wireless communication module 104, memory module 112, invisible light signal transmitter 108, and short-range wireless communication module 106, directly or indirectly via bus 114.

Fig. 2 shows a block diagram of input, control and output hardware modules in a receiving device 200 according to an embodiment of the application. Referring to fig. 2, the receiving device 200 includes a visible light camera module 202, a short range wireless communication module 204, a touch screen module 206, a processor module 208, a memory module 210, and a bus 212. The receiving device 200 may be a terminal such as a mobile phone, a smart phone, a tablet computer, a laptop computer or a desktop computer.

The visible light camera module 202 is an input hardware module and is used to acquire frames when the receiving method 350 in fig. 3 is executed, when the exposure time interval reduction method 1000 in fig. 10 is executed, when the color conversion method 1300 in fig. 13 is executed, when the auto-focusing method 1500 in fig. 15 is executed, and when the method 1800 of auto-focusing in the case of multiple transmitting devices in fig. 18 is executed. The receiving method 350, the exposure time interval reduction method 1000, the color conversion method 1300, the auto focus method 1500, and the method 1800 of auto focus in the case of a plurality of transmitting apparatuses are executed by the receiving apparatus 200. At least a portion of the frame is to be sent to the processor module 204. The visible light camera module 202 can capture the invisible light emitted from the invisible light signal emitter 108 in fig. 1. The visible light camera module 202 may be a color camera such as an RGB camera. Alternatively, another color camera may also be used as the visible light camera module 202, such as a YUV camera or Bayer raw data camera (Bayer raw camera). The visible light camera module 202 includes an image sensor having a two-dimensional Photodiode (PD) array structure. The image sensor may be a charge-coupled device (CCD) image sensor or a complementary metal-oxide semiconductor (CMOS) image sensor.

The short-range wireless communication module 204 may serve as an input hardware module and an output hardware module. The short range wireless communication module 204 performs the step of broadcasting an advertisement message, the step of transmitting a pairing request, the step of receiving a pairing request response, and the step of pairing with the receiving device 200 in the receiving method 350 using a short range wireless communication technology such as Bluetooth Low Energy (BLE). For example, the advertisement message is received from the processor module 208. Upon receiving the instruction from the processor module 208, the short-range wireless communication module 204 uses BLE for communication. Alternatively, another short-range wireless communication technology, such as ZigBee, may be used.

The touch screen module 206 may serve as an input hardware module and an output hardware module. The touch screen module 206 is used to display frames acquired by the visible light camera module 202 when the receiving method 350 is executed, when the exposure time interval reduction method 1000 is executed, when the color conversion method 1300 is executed, when the auto focus method 1500 is executed, and when the method 1800 of auto focus is executed with multiple transmitting devices. The touch screen module 206 displays the frames when receiving instructions directly or indirectly from the processor module 204. The touch screen module 206 is also used to receive touch input from a user when the method 1800 of autofocusing with multiple sending devices is performed. The touch input may be communicated to the processor module 204, or other module controlled by the processor module 204, for further processing.

The memory module 210 may be a transitory or non-transitory computer readable medium comprising at least one flash stored program instruction that, when executed by the processor module 208, causes the processor module 208 to control the visible light camera module 202, the short range wireless communication module 204, and the touch screen module 206 to perform the receiving method 350, the exposure time interval reduction method 1000, the color conversion method 1300, the auto-focus method 1500, and the method 1800 of auto-focus in the case of multiple transmitting devices. The processor module 208 includes at least one processor that sends signals directly or indirectly to the visible light camera module 202, the short range wireless communication module 204, the touch screen module 206, and the memory module 210 over the bus 212 and/or receives signals directly or indirectly from the visible light camera module 202, the short range wireless communication module 204, the touch screen module 206, and the memory module 210 over the bus 150. The at least one processor may be a Central Processing Unit (CPU), a Graphics Processing Unit (GPU), and/or a Digital Signal Processor (DSP). the CPU may send the fragments of the frame, some program instructions, and other data or instructions to the GPU and/or DSP via bus 212.

The receiving device 200 is a type of computing system that has all of its components integrated together via a bus 212. Other types of computing systems, such as a computer system having a remote camera module instead of the camera module 202, are also within the intended scope of the present application.

Fig. 3 shows a flowchart of a communication method of a transmitting device and a receiving device according to an embodiment of the present application. Referring to fig. 3, the communication method includes a transmission method 300 and a reception method 350. The transmission method 300 includes the steps on the left side of fig. 3. The receiving method 350 comprises the steps of the right side of fig. 3. In step 302, an encoded signal pulse pattern having a data sequence is loaded. In step 352, the camera frame rate is broadcast. In step 304, the camera frame rate broadcast by the receiving device is scanned. In step 306, the first invisible light signal emitter emits a first signal pulse at a signal pulse rate substantially the same as the camera frame rate. The signal pulse pattern is repeatedly transmitted in said first signal pulse. In step 354, the visible light camera module captures a plurality of first frames at the camera frame rate. Each first frame includes a respective first region generated by a respective one of the first portions of the first signal pulses. In step 356, the corresponding first region in each first frame is detected. In step 358, the corresponding first region in each first frame is marked to obtain a first coding pattern corresponding to the signal pulse pattern, and the first coding pattern is decoded to identify the data sequence. In step 360, a pairing request including the data sequence is transmitted to the transmitting device. In step 308, the pairing request is received from the receiving device. In step 310, a pairing request response is sent to the receiving device. In step 362, the pairing request response is received from the sending device. In step 364, the receiving device is paired with the transmitting device based on the data sequence. In step 312, the transmitting device is paired with the receiving device based on the data sequence.

The term "substantially" or "essentially" refers to a variation in the number that a skilled person can expect with regard to the precision with which the transmitting device and the receiving device are used.

The transmission method 300 is performed by the transmission apparatus 100 in fig. 1. The receiving method 350 is performed by the receiving apparatus 200 in fig. 2.

In step 302, an encoded signal pulse pattern having a data sequence is loaded. The data sequence may be encoded into a signal pulse pattern by Pulse Width Modulation (PWM). The signal pulse pattern may be divided into a plurality of first coding units. Each first coding unit corresponds to a bit of the data sequence and comprises a plurality of bright and dark signal pulses. If the bright portion of the bright-dark signal pulse is less than 50% (e.g., 40%) for each of the second coding units in the first coding unit, the bit of the data sequence corresponding to each of the second coding units is "0". If the bright portion of the bright-dark signal pulse is more than 50% (e.g., 60%) for each of the third encoding units in the first encoding unit, the bit of the data sequence corresponding to each of the third encoding units is "1". Alternatively, another percentage threshold may be used, for example 40% or 60%, below which a coding unit corresponds to a "0" and above which a coding unit corresponds to a "1".

In step 352, the camera frame rate is broadcast. In step 304, the camera frame rate broadcast by the receiving device is scanned. Typically, the visible light camera module 202 has several options for the camera frame rate, such as 30 frames per second (fps) or 60 fps. The broadcasted camera frame rate is the camera frame rate used to acquire frames when the receiving method 350 is executed, when the exposure time interval reduction method 1000 in fig. 10 is executed, when the color conversion method 1300 in fig. 13 is executed, when the auto focus method 1500 in fig. 15 is executed, and when the method 1800 of auto focus in the case of multiple transmitting devices in fig. 18 is executed. Steps 352 and 304 are a first handshake procedure between the transmitting device 100 and the receiving device 200 using the short range wireless communication technology.

In step 306, the first invisible light signal emitter emits a first signal pulse at a signal pulse rate substantially the same as the camera frame rate. The signal pulse pattern is repeatedly transmitted in said first signal pulse. In step 354, the visible light camera module captures a plurality of first frames at the camera frame rate. Each first frame includes a respective first region generated by a respective one of the first portions of the first signal pulses. The first invisible light signal emitter may be the invisible light signal emitter 108 in fig. 1. The visible light camera module may be the visible light camera module 202 in fig. 2. Fig. 4 shows a schematic diagram of one 402 of a plurality of first frames captured by the visible light camera module 202 of the receiving apparatus 200 according to an embodiment of the present application. In the example of fig. 4, the transmitting device 100 is a smart speaker. The receiving device 200 is a mobile phone 406. The image of the smart speaker is captured in each first frame, and there is a corresponding region in each first frame (e.g., region 404 in one 402 of the plurality of first frames). An image of the invisible light signal emitter 108 of the smart speaker is captured in each first frame, and there is a corresponding area (e.g., area 408 in area 404) on the corresponding area of the smart speaker in each first frame. By pairing using these first frames, the effect of "what you see is what you get" is achieved. In the related art, there is a list of transmitting devices based on text-based identification codes (IDs), and a user selects one of the transmitting devices to pair, which is more intuitive in the manner of using the first frames for pairing.

Fig. 5 shows a timing diagram of a synchronization start timing of a first frame at a signal pulse rate substantially the same as the camera frame rate for a first portion of a first signal pulse and for which sampling is successful, according to an embodiment of the application. In the example shown in fig. 5, the invisible light signal emitter 108 has a signal pulse interval I502 based on the signal pulse rate. The signal pulse interval 502 corresponds to a first portion of a first signal pulse. The visible light camera module 202 has a sampling clock period I504 based on the camera frame rate. The sampling clock period 504 corresponds to the first frame. The shutter of the visible light camera module 202 has an exposure time interval I504 that is half of the sampling clock period I502. The signal pulse interval I502 is substantially the same as the sampling clock period I502. The signal pulse pattern is repeatedly transmitted in said first signal pulse. In the example shown in fig. 5, there are 5 signal pulses in one of the first coding units. The bright signal pulses are represented by a lower density dot in one box in the signal pulse interval. Dark signal pulses are represented by a higher density of dots in one box in the signal pulse interval. When there are 2 bright signal pulses out of the 5 signal pulses, the bit of the data sequence corresponding to the 5 signal pulses is "0". When there are 3 bright signal pulses out of the 5 signal pulses, the bit of the data sequence corresponding to the 5 signal pulses is "1". Assume that the first bit of the data sequence is a "0". Each first frame is captured during the exposure time of a corresponding one of the plurality of sampling clock cycles 504. There is a first portion of the first signal pulse and a synchronized start timing of the first frame between the non-visible light signal emitter 108 and the visible light camera module 202. That is, both the signal pulse interval 502 and the sampling clock period 504 begin at time T502. The exposure time of each sampling clock cycle 504 is within a corresponding one of the plurality of signal pulse intervals 502. Therefore, sampling can be successful. The same applies to the case where the first bit of the data sequence is "1", as shown in the lower side of fig. 5.

Fig. 6 shows a timing diagram of the signal pulse rate substantially the same as the camera frame rate, the start timing for the first portion of the first signal pulse and the first frame to make the sampling successful in a first asynchronous manner, according to an embodiment of the present application. Compared to the first part of the first signal pulse and the start timing of the first frame in the example shown in fig. 5, the first part of the first signal pulse and the start timing of the first frame in the example shown in fig. 6 belong to a first asynchronous manner. That is, the signal pulse interval 602 begins at T602, and the sampling clock period 604 begins at T604, which is later than T602 but within the first signal pulse interval 602. The exposure time of each sampling clock cycle 604 is within the corresponding signal pulse interval of the plurality of signal pulse intervals 602. Therefore, sampling can be successful. The same applies to the case where the first bit of the data sequence is "1", as shown in the lower side of fig. 6.

Fig. 7 shows a timing diagram of the start timing of a first frame at a signal pulse rate substantially the same as the camera frame rate for a first portion of a first signal pulse and for sampling to be successful in a second asynchronous manner, according to an embodiment of the application. Compared to the first part of the first signal pulse and the start timing of the first frame in the example shown in fig. 5, the first part of the first signal pulse and the start timing of the first frame in the example shown in fig. 7 belong to the second asynchronous manner. That is, the signal pulse interval 702 begins at T702, and the sampling clock period 704 begins at T704, which is later than T702 but within the second signal pulse interval 702. The exposure time of each sampling clock cycle 704 is within the corresponding signal pulse interval of the plurality of signal pulse intervals 702. Since there are 2 bright signal pulses in the second through sixth signal pulses in the first portion of the first signal pulse and the signal pulse pattern is repeatedly transmitted in the first portion of the first signal pulse, the second through sixth signal pulses in the first portion of the first signal pulse may still represent a case where the first bit of the data sequence is "0". Therefore, sampling can be successful. The same applies to the case where the first bit of the data sequence is "1", as shown in the lower side of fig. 7.

Fig. 8 shows a timing diagram of the signal pulse rate substantially the same as the camera frame rate, for a first portion of a first signal pulse and the start timing of a first frame in a third asynchronous manner to make the sampling successful according to an embodiment of the application. The first portion of the first signal pulse and the start timing of the first frame in the example shown in fig. 8 belong to a third asynchronous manner, compared to the first portion of the first signal pulse and the start timing of the first frame in the example shown in fig. 5. That is, the signal pulse interval 802 begins at T802, and the sampling clock period 804 begins at T804, which is later than T802 but within the fourth signal pulse interval 802. The exposure time of each sampling clock cycle 804 is within a corresponding one of the plurality of signal pulse intervals 802. Because the fourth and fifth signal pulses in the first portion of the first signal pulse are dark, the fourth and fifth signal pulses in the first portion of the first signal pulse are discarded. The sixth to tenth signal pulses in the first part of the first signal pulse correspond to the second bit of the first acquired data sequence. Since the signal pulse pattern is repeatedly transmitted in the first signal pulse, the signal pulse corresponding to the first bit of the data sequence is finally obtained. Therefore, sampling can be successful. The same applies to the case where the first bit of the data sequence is "1", as shown in the lower side of fig. 8.

FIG. 9 shows a timing diagram of the start timing of a first frame at a signal pulse rate substantially the same as the camera frame rate, for a first portion of a first signal pulse, and for which sampling may be unsuccessful in a fourth asynchronous manner, according to an embodiment of the application. Compared to the first portion of the first signal pulse and the start timing of the first frame in the example shown in fig. 5, the first portion of the first signal pulse and the start timing of the first frame in the example shown in fig. 9 belong to a fourth asynchronous manner. That is, the signal pulse interval 902 begins at T902 and the sampling clock period 904 begins at T904, which is later than T902 but within the first signal pulse interval 902. However, the exposure time of one of the plurality of sampling clock periods 904 is across two of the plurality of signal pulse intervals 902, wherein one signal pulse interval corresponds to a bright signal pulse and the other signal pulse interval corresponds to a dark signal pulse, e.g., the second and third signal pulse intervals in the first portion of the first signal pulse, which causes the sampled bright or dark signal pulse to be confused. Therefore, sampling was unsuccessful. The same applies to the case where the first bit of the data sequence is "1", as shown in the lower side of fig. 9.

Fig. 10 shows a flow chart of an exposure time interval reduction method 1000 according to an embodiment of the present application to improve the unsuccessful sampling condition of fig. 9 and to improve the visibility of the portion of the first frame corresponding to the first region of the bright signal pulses of the first portion of the first signal pulses. Referring to fig. 10, the receiving method 350 further includes an exposure time interval reduction method 1000 between step 352 and step 354.

In step 1002, the visibility of the overexposed second region in the second frame captured at the exposure time interval of the shutter of the visible camera module is obtained. In step 1004, the exposure time interval is decreased such that the visibility of the third region of the third frame captured at the decreased exposure time interval is higher than the visibility of the second region. In step 1002, the visibility of the overexposed second region in the second frame captured at the exposure time interval of the shutter of the visible camera module is obtained. In step 1004, the exposure time interval is decreased such that the visibility of the third region of the third frame captured at the decreased exposure time interval is higher than the visibility of the second region. Fig. 11 shows a schematic diagram 1100 of exemplary photo-electric characteristic curves 1102, 1108 and 1114 for different exposure time intervals E1, E2 and E3, wherein a reduced exposure time interval E2 or E3 improves visibility of a first region of a first frame of bright signal pulses corresponding to a first portion of the first signal pulse, in accordance with an embodiment of the present application. The X-axis of the graph 1100 represents the light intensity captured by a pixel of the visible-light camera 202, and the Y-axis of the graph 1100 represents the pixel output voltage of the pixel of the visible-light camera 202. In step 1002, the exposure time interval, e.g., exposure time interval E1, is equal to one-half of the sampling clock period I502 (as shown in fig. 5). Since the invisible light signal emitter 108 emits invisible light, the second area may be generated by a bright signal pulse of the first signal pulse captured in the second frame. In the case of the exposure time interval E1, the pixels of the visible light camera 202 correspond to the photoelectric characteristic curve 1102. The photoelectric characteristic 1102 has a first section and a second section. The pixel output voltage decreasing along the first segment of the photo characteristic 1102 slowly decreases as the light intensity decreases compared to the pixel output voltage decreasing along the second segment of the photo characteristic 1102. In the example of fig. 11, the slope of the first segment of the photoelectric characteristic curve 1102 is 0, and the slope of the second segment of the photoelectric characteristic curve 1102 is greater than the slope of the first segment of the photoelectric characteristic curve 1102. The point 1104 is located on the first segment of the photoelectric characteristic 1102 and is not at the boundary between the first segment of the photoelectric characteristic 1102 and the second segment of the photoelectric characteristic 1102. Point 1104 represents the pixel output voltage of the pixel of the visible camera 202 reaching the saturation level SL as a result of the intensity of light generated by the invisible light signal emitter 108 during the exposure time interval E1. Therefore, pixels in the second area corresponding to the pixels of the visible-light camera 202 are overexposed. A point 1106 on the photo characteristics curve 1102 indicates that the pixel output voltage of the pixels of the visible light camera 202 has reached a level L1 as a result of the intensity of light generated by the visible light at the exposure time interval E1. For example, the visible light is reflected from the first portion of the transmitting device 100.

In step 1004, the exposure time interval E1 is decreased until it reaches exposure time interval E2. The third region may be generated by a bright signal pulse among the first signal pulses captured in the third frame at exposure time interval E2. In the case of the exposure time interval E2, the pixels of the visible light camera 202 correspond to the photoelectric characteristic curve 1108. The optoelectronic characteristic 1108 has a first section and a second section. The pixel output voltage decreasing along the first segment of the photoelectric characteristic curve 1108 decreases slowly with decreasing light intensity as compared to the pixel output voltage decreasing along the second segment of the photoelectric characteristic curve 1108. Furthermore, the pixel output voltage increasing along the second segment of the PV characteristic curve 1108 increases slowly with increasing light intensity as compared to the pixel output voltage increasing along the second segment of the PV characteristic curve 1102. In the example of FIG. 11, the slope of the first segment of the PV characteristic curve 1108 is 0, and the slope of the second segment of the PV characteristic curve 1108 is greater than the slope of the first segment of the PV characteristic curve 1108 but less than the slope of the second segment of the PV characteristic curve 1102. The point 1110 is located on the first section of the photoelectric characteristic curve 1108 and is not at the boundary between the first section of the photoelectric characteristic curve 1108 and the second section of the photoelectric characteristic curve 1108. Point 1110 represents the pixel output voltage of the pixel of the visible camera 202 reaching the saturation level SL as a result of the intensity of light generated by the invisible light signal emitter 108 for the exposure time interval E2. Therefore, pixels in the third area corresponding to the pixels of the visible-light camera 202 are overexposed. A point 1112 on the electro-optic characteristic curve 1108 represents the pixel output voltage of the pixels of the visible-light camera 202 reaching the level L2 as a result of the intensity of light generated by the visible light at the exposure time interval E2. For example, the visible light is reflected from the first portion of the transmitting device 100. Since the exposure time interval E2 is shorter than the exposure time interval E1 and the slope of the second segment of the photoelectric characteristic curve 1108 is smaller than the slope of the second segment of the photoelectric characteristic curve 1102, the level L2 is lower than the level L1. Since the level L2 is farther from the saturation level SL than the level L1, the visibility of the third region is higher than that of the second region. The capturing of the first frame by the exposure time interval E2 improves the visibility of the portion of the first area of the bright signal pulse in the first frame that corresponds to the first portion of the first signal pulse.

In step 1004, the exposure time interval E1 may be further decreased until it reaches exposure time interval E3. The third region may be generated by a bright signal pulse among the first signal pulses captured in the third frame at exposure time interval E3. In the case of the exposure time interval E3, the pixels of the visible light camera 202 correspond to the photoelectric characteristic curve 1114. The photoelectric characteristic curve 1114 has a first section and a second section. The pixel output voltage decreasing along the first segment of the photoelectric characteristic curve 1114 decreases slowly as the light intensity decreases compared to the pixel output voltage decreasing along the second segment of the photoelectric characteristic curve 1114. Furthermore, the pixel output voltage increasing along the second segment of the electro-optic characteristic curve 1114 increases slowly with increasing light intensity as compared to the pixel output voltage increasing along the second segment of the electro-optic characteristic curve 1108. In the example of fig. 11, the slope of the first segment of the optical-electrical characteristic curve 1114 is 0, and the slope of the second segment of the optical-electrical characteristic curve 1114 is greater than the slope of the first segment of the optical-electrical characteristic curve 1114 but less than the slope of the second segment of the optical-electrical characteristic curve 1108. The point 1116 is located at a boundary between the first segment of the electro-optic characteristic 1114 and the second segment of the electro-optic characteristic 1114. Point 1116 represents the pixel output voltage of the pixel of the visible light camera 202 having reached the saturation level SL as a result of the intensity of light generated by the invisible light signal emitter 108 during exposure time interval E3. A point 1118 on the photoelectric characteristic curve 1114 represents the pixel output voltage of the pixels of the visible-light camera 202 induced by the light intensity generated by the visible light at the exposure time interval E3 reaching the level L3. For example, the visible light is reflected from the first portion of the transmitting device 100. Since the exposure time interval E3 is shorter than the exposure time interval E2 and the slope of the second section of the photoelectric characteristic curve 1114 is smaller than the slope of the second section of the photoelectric characteristic curve 1108, the level L3 is lower than the level L2. Since the level L3 is farther from the saturation level SL than the level L2, the visibility of the third region in the exposure time interval E3 is higher than the visibility of the third region in the exposure time interval E2. In addition, because the point 1116 is located at the boundary between the first section of the photoelectric characteristic curve 1114 and the second section of the photoelectric characteristic curve 1114, further lowering of the exposure time interval E3 may cause the pixel output voltage of the pixels of the visible-light camera 202 to become a level lower than the saturation level, so that the visibility of the third region captured at this exposure time interval is lower than the visibility of the third region captured at the exposure time interval E3. Therefore, the visibility of the third area captured under the exposure time interval E3 is the highest. The pixels in the third area captured under the exposure time interval E3 are properly exposed. The first frame, in which the visibility of the portion of the first area of the bright signal pulse corresponding to the first portion of the first signal pulse is highest, is captured by exposure time interval E3.

In the example of fig. 11, the first and second sections of the photoelectric characteristic curves 1102, 1108, and 1114 are straight lines. Other forms of optoelectronic characteristic, such as those comprising logarithmic sections, are also within the intended scope of the present application.

FIG. 12 shows a timing diagram of a reduced exposure time interval to improve the sampling failure condition of FIG. 9, according to an embodiment of the present application. The exposure time interval I1202 in the example of fig. 12 is the reduced exposure time interval of fig. 10 compared to the exposure time interval I506 (shown in fig. 5) in the example of fig. 9. Because the exposure time interval I1202 is shorter than the exposure time interval I506, the exposure time of one sampling clock cycle 1204 has less chance to span two signal pulse intervals in the signal pulse interval 1202, one of which corresponds to a bright signal pulse and the other of which corresponds to a dark signal pulse. For example, the exposure time of exposure time interval I506 spans the second and third pulses in the first portion of the first signal pulse. The exposure time of exposure time interval I1202 is within the second pulse in the first portion of the first signal pulse, avoiding the sampling of a mixed line of whether the second pulse or the third pulse is in the first portion of the first signal pulse. For the shortest exposure time E3 described with reference to fig. 11, the exposure time of one sampling clock cycle 1204 has the least chance of spanning two signal pulse intervals 1202, one of which corresponds to a bright signal pulse and the other of which corresponds to a dark signal pulse.

Fig. 13 is a flowchart illustrating a color conversion method 1300 according to an embodiment of the present disclosure, in which the color conversion method 1300 makes a portion of a first region more distinguishable from other regions of a first frame, and if the color conversion method 1300 is not executed, the other portions of the first frame have a color similar to the portion of the first region. Referring to fig. 13, the receiving method 350 further includes a color conversion method 1300 between step 352 and step 354. In step 1302, the white balance of the visible light camera module is adjusted such that the color of the invisible light emitted by the first invisible light signal emitter is more distinguishable from the color of visible light, which is similar to the color of the invisible light emitted by the first invisible light signal emitter at neutral white balance.

In step 1302, the white balance of the visible light camera module is adjusted such that the color of the invisible light emitted by the first invisible light signal emitter is more distinguishable from the color of visible light, which is similar to the color of the invisible light emitted by the first invisible light signal emitter at neutral white balance. FIG. 14 shows a schematic diagram of an exemplary color conversion scene according to an embodiment of the present application. In the example of fig. 14, if the visible camera module 202 (shown in fig. 2) captures the first frame 1402 at neutral white balance including a region 1404 and a region 1406, the region 1404 corresponding to the outer cover of the smart speaker surrounding the non-visible light signal emitter 108 and the region 1406 corresponding to the non-visible light signal emitter 108, the region 1404 has a higher chance of being colored similar to the region 1406, as shown, the region 1404 and the region 1406 having the same fill pattern. The reason is that most indoor lighting causes the visible light reflected in the indoor environment captured in a frame at neutral white balance to become warmer. When the cover of the smart speaker is colored white, for example, room lighting may cause visible light reflected from the cover of the smart speaker that is captured in frame 1402 at neutral white balance to become warm. In embodiments where the invisible light signal emitter 108 emits infrared light, the color of the infrared light captured in the frame 1402 also appears warm. In step 1302, the white balance of the visible light camera module 202 is adjusted to a color temperature, such that the frames captured by the visible light camera module 202 with the adjusted white balance are focused to a blue color. In one embodiment, the white balance is adjusted to a color temperature of less than 3000K. In this case, the red light appears gray in the frame, and the infrared light appears grayish-purple in the frame. When the visible camera module 202 captures a frame 1408 including a region 1410 and a region 1412 at the adjusted white balance, the region 1410 corresponding to the cover of the smart speaker around the non-visible light signal emitter 108 and the region 1412 corresponding to the non-visible light signal emitter 108, then the region 1410 has a smaller chance of having a color similar to the region 1412, as shown, the region 1410 is filled with a fill pattern and the region 1412 is not filled with a fill pattern. By using the adjusted white balance to capture the first frame, the portion of the first region is more distinguishable from other regions of the first frame, which would be similar to the portion of the first region if the color conversion method 1300 were not performed.

In the example of fig. 14, the invisible light signal emitter 108 emits infrared light. Alternatively, the invisible light signal emitter emits ultraviolet light, and the visible light camera module 202 captures the ultraviolet light, and the white balance of the visible light camera module 202 is adjusted to a certain color temperature, so that the frame captured by the visible light camera module 202 with the adjusted white balance is focused to a red tone.

Fig. 15 shows a flow chart of an auto-focusing method 1500 according to an embodiment of the present application, the method 1500 being used for auto-focusing on a first invisible light signal emitter when the first invisible light signal emitter emits a first portion of a first signal pulse. Referring to fig. 15, the receiving method 350 further includes a method 1500 between step 352 and step 354. In step 1502, a plurality of fourth frames are captured at the camera frame rate, wherein each fourth frame includes a respective fourth region generated by a respective signal pulse in the second portion of the first signal pulse and a respective fifth region corresponding to a portion of the transmitting device that does not include the first invisible light signal emitter. After step 1502, step 1002 is performed. And, in step 1504, aligning the respective fourth regions in each fourth frame with each other by aligning the respective fifth regions in each fourth frame with each other. In step 1506, the corresponding fourth region in each fourth frame is detected to determine a location of the corresponding fourth region. In step 1508, the first invisible light signal emitter is identified as an object on the object tracking by the location of each of at least one of the plurality of fourth areas to automatically focus on the first invisible light signal emitter when the first invisible light signal emitter emits the first portion of the first signal pulse.

FIG. 16 is a diagram illustrating the acquisition, alignment, and detection steps 1502,1504, and 1506 of the auto-focusing method 1500 shown in FIG. 15, according to an embodiment of the present application. Referring to fig. 16, in step 1502, a plurality of fourth frames (e.g., 1602 and 1604) are captured at the camera frame rate, wherein each fourth frame 1602 or 1604 includes a respective fourth region 1606 or 1608 and a respective fifth region 1610 or 1612, the respective fourth region 1606 or 1608 being generated by a respective signal pulse in the second portion of the first signal pulse, and the respective fifth region 1610 or 1612 corresponding to a portion of the transmitting device 100 (shown in fig. 1) that does not include the first invisible light signal emitter 108. In step 1504, the corresponding fourth regions 1606 or 1608 (and 1608 or 1606) in each fourth frame 1602 or 1604 are aligned with each other (as indicated by the dashed lines of the fourth regions 1606 and 1608) by aligning the corresponding fifth regions 1610 or 1612 (and 1612 or 1610) in each fourth frame 1602 or 1604 with each other. In one embodiment, the texture of the fifth region 1610 or 1612 may be used in the alignment. After step 1504, fourth areas 1606 and 1608 are aligned in the up-down direction and the left-right direction. For the sake of simplicity, the alignment in the up-down direction is not shown. In step 1506, the corresponding fourth region 1606 or 1608 in each fourth frame 1602 or 1604 is detected to obtain a location of the corresponding fourth region 1606 or 1608. The fourth regions 1606 and 1608 can be detected based on a characteristic that the region at the aligned position flickers. The location of the last region 1608 of the fourth regions 1606 and 1608 may be used in autofocus. Alternatively, all locations of fourth regions 1606 and 1608 are used for autofocus.

In the example of FIG. 16, two fourth regions 1606 and 1608 are shown. Other numbers of fourth regions are also within the intended scope of the present application.

Fig. 17 is a diagram illustrating an auto-focus step 1508 of the auto-focus method 1500 shown in fig. 15 according to an embodiment of the present application. Referring to fig. 17, in step 1508, the first invisible light signal emitter 108 is identified as an object on the object tracking by the position of each of at least one of the fourth regions 1606 and 1608 (shown in fig. 16) of the plurality of fourth regions to automatically focus on the first invisible light signal emitter 108 when the first invisible light signal emitter 108 emits the first portion of the first signal pulse. A first portion of the first signal pulse is acquired in a first frame 1702 while the first invisible light signal emitter 108 emits the first portion of the first signal pulse. By identifying the first invisible light signal emitter 108 as an object on object tracking, autofocus remains on the invisible light signal emitter 108 even if a user holding the receiving device 200 (shown in fig. 2) moves his or her hand while capturing the first frame 1702. This portion is present in the first frame 1702 in first regions 1704.. 1706, which correspond to the first portion of the first signal pulse and are enclosed by dashed lines, even if these first regions 1704.. 1706 are at different locations in the corresponding frame in the first frame 1702.

Fig. 18 shows a flow diagram of a method 1800 of autofocusing in the case of multiple sending devices according to an embodiment of the application. Method 1800 reinforces the case where method 1500 encounters multiple transmitting devices. Referring to fig. 18, the receiving method 350 further includes a method 1800 between step 352 and step 354. In step 1802, a plurality of fourth frames are captured at the camera frame rate, wherein each fourth frame includes a respective fourth region generated from a respective signal pulse in the second portion of the first signal pulse, at least one respective fifth region generated from a respective signal pulse in the second signal pulse emitted by the second invisible light signal emitter of a respective one of the at least one other transmitting device, and a respective sixth region corresponding to a portion of the transmitting device that does not include the first invisible light signal emitter or a portion of any of the at least one other transmitting device that does not include the second invisible light signal emitter. In step 1804, the respective fourth regions in each fourth frame are aligned with each other by aligning the respective sixth regions with each other in each fourth frame. In step 1806, the corresponding fourth region and the at least one corresponding fifth region in each fourth frame are detected to obtain a location of the corresponding fourth region and a location of each of the at least one corresponding fifth region. In step 1808, the number of the fourth region and the at least one fifth region in a fourth frame is calculated to obtain the number of the transmitting device and the at least one other transmitting device. In step 1810, for each of the transmitting device and the at least one other transmitting device, the region sizes of a plurality of corresponding regions in the fourth frame are averaged. In step 1812, the position of each of the at least one fourth region in the plurality of fourth regions is selected to perform auto-focusing for the largest one of the fourth region and the at least one fifth region in a fourth frame of the plurality of fourth frames based on the region size of the fourth region in the fourth frame of the plurality of fourth frames. In step 1804, it is determined whether the user touch operation changes the selection. If so, then step 1816 is performed. If not, then step 1508 is performed. In step 1816, autofocus is performed based on the changed selection.

FIG. 19 shows a schematic diagram of an exemplary multi-send device scenario to which the auto-focus method 1800 of FIG. 18 may be applied, according to an embodiment of the present application. Steps 1802, 1804 and 1806 are similar to steps 1502,1504 and 1506 shown in fig. 16, so only one fourth frame is shown in fig. 19 for simplicity. Referring to fig. 19, in step 1802, a plurality of fourth frames (one of which 1902 is illustratively shown) are captured at the camera frame rate, wherein each fourth frame 1902 includes a respective fourth region 1904, at least one respective fifth region (one of the fifth regions 1906 being exemplarily shown) and a respective sixth region 1908 or 1910, the respective fourth region 1904 is generated by a respective signal pulse in the second portion of the first signal pulse, each respective fifth region 1906 is generated by a respective one of the second signal pulses emitted by the second invisible light signal emitter of a respective one of the at least one other transmitting device, and the corresponding sixth region 1908 or 1910 corresponds to a portion of the transmitting device 100 (shown in figure 1) that does not include the first invisible light signal emitter 108, or to a portion of any of the at least one other transmitting devices that does not include the second invisible light signal emitter. The at least one corresponding fifth area corresponds to the at least one other transmitting device. The second signal pulse corresponds to the fourth frame 1902. In the example of fig. 19, the transmitting device 100 is a smart speaker. The at least one other transmitting device is a smart television. In step 1804, the corresponding fourth areas 1904 in each fourth frame 1902 are aligned with each other by aligning the corresponding sixth areas 1908 or 1910 in each fourth frame 1902 with each other. In step 1806, the corresponding fourth region 1904 and the at least one corresponding fifth region 1906 in each fourth frame 1902 are detected to obtain a location of the corresponding fourth region 1904 and a location of each of the at least one corresponding fifth region 1906. In step 1808, the number of the fourth region 1904 and the at least one fifth region 1906 in a fourth frame 1902 is calculated to obtain the number of the sending device 100 and the at least one other sending device. In the example of fig. 19, the number of the fourth area 1904 and the at least one fifth area 1906 is two, which means that the number of the transmission apparatus 100 and the at least one other transmission apparatus is two. In step 1810, based on the number of the transmitting device 100 and the at least one other transmitting device, the region sizes of the plurality of corresponding regions in the fourth frame 1902 are averaged for each of the transmitting device 100 and the at least one other transmitting device. Each respective region is one of the fourth region 1904 and the at least one fifth region 1906 in a respective one of a plurality of fourth frames 1902. The one of the fourth region 1904 and the at least one fifth region 1906 is generated by a bright signal pulse of the first signal pulses or by a bright signal pulse of the second signal pulses corresponding to one of the at least one fifth region 1906. In step 1812, the position of each of at least one of the plurality of fourth regions 1904 is selected to autofocus for the largest of the fourth region 1904 and the at least one fifth region 1906 in a fourth frame 1902 based on the size of the fourth region 1904 in the fourth frame 1902. In step 1804, it is determined whether the user touch operation is to change the selection. The user may select the smart tv by touch. If so, then step 1816 is performed. If not, then step 1508 is performed. In step 1816, autofocus is performed based on the changed selection.

Referring to fig. 3, a receiving method 350 includes a method of analyzing a first frame to obtain a data sequence, the method including steps 356 and 358. In step 356, the corresponding first region in each first frame is detected. In step 358, the corresponding first region in each first frame is marked to obtain a first coding pattern corresponding to the signal pulse pattern, and the first coding pattern is decoded to identify the data sequence. Fig. 20 shows a flow chart of the detection step 356 in the method of analyzing the first frame to obtain a data sequence according to an embodiment of the application. Step 356 includes the following steps. Referring to fig. 20, in step 2002, each first frame is binarized using a threshold value to obtain a binarized frame. In step 2004, at least one respective first bright region is extracted from each binarized frame to obtain a location of each of the at least one respective first bright regions. In step 2006, the location of each of the at least one respective first bright regions is tracked to detect each of at least one set of bright and dark regions corresponding to the binarized frame and being location-dependent, wherein the at least one set of bright and dark regions includes the first region.

Fig. 21 shows a schematic diagram of the binarization step 2002 of removing the region that is not the first region based on the luminance or the color in the detection step 356 according to the embodiment of the present application. In step 2002, each first frame 2102 is binarized using a threshold value to obtain binarized frames (an exemplary display of one of binarized frames 2104 corresponding to the first frames 2102). In one embodiment, the threshold is within a proximity of at least one pixel value of a pixel in one of the first regions (one of the first regions 2106 is illustratively shown). The first region 2106 of the plurality of first regions corresponds to a bright signal pulse of the first signal pulses. The at least one pixel value is a luminance value (e.g., brightness or luminance). In the example of fig. 21, since the first area 2106 of the plurality of first areas corresponds to the bright signal pulse of the first signal pulse of the invisible light signal emitter 108, and the areas 2108 and 2110 correspond to the circular area and the square area of the smart tv display, the first area 2106 and the areas 2108 and 2110 of the plurality of first areas have higher luminance values than other portions of the first frame 2102. Further, since the threshold value is set to be within the vicinity of the luminance value of one pixel in the first region 2106 among the plurality of first regions, the regions 2112, 2114, and 2116 corresponding to the first region 2106 and the regions 2108 and 2110 among the plurality of first regions are bright regions in the binarized frame 2104. The regions 2112, 2114, and 2116 are candidate regions for the first region 2106 of the plurality of first regions to be used to identify a data sequence. Other parts of the first frame 2102 are removed to find candidate regions in the binarized frame 2104.

Alternatively, where the color conversion method 1300 has been performed, the threshold is within a neighborhood of a set of pixel values for a pixel in one of the plurality of first regions. The first region 2106 of the plurality of first regions corresponds to a bright signal pulse of the first signal pulses. The set of pixel values includes a color value. In one embodiment, the set of pixel values are RGB color values. The smart speaker in the example of fig. 21 is similar to the smart speaker in the example of fig. 14. Without performing the color conversion method 1300, the color of the region 2118 corresponding to the cover of the smart speaker surrounding the invisible light signal emitter 108 may be difficult to distinguish from the color of the first region 2106 of the plurality of first regions corresponding to the invisible light signal emitter 108. By performing the color conversion method 1300 and using a threshold value in the vicinity of the RGB color value of a pixel in a first one of the plurality of first regions, the region 2112 corresponding to the first one of the plurality of first regions 2106 is a bright region in the binarized frame 2104 and a region (not labeled) corresponding to the region 2118 is a dark region in the binarized frame 2104. Further, after the color conversion method 1300 is performed, the colors of the areas 2108 and 2110 corresponding to the circular area and the square area of the smart tv display are similar to the first area 2106 of the plurality of first areas. Thus, regions 2114 and 2116 corresponding to the regions 2108 and 2110 are also bright regions in the binarized frame 2104. The regions 2112, 2114, and 2116 are candidate regions for the first region 2106 of the plurality of first regions to be used to identify a data sequence. Other parts of the first frame 2102 are removed to find candidate regions in the binarized frame 2104.

Fig. 22 shows a schematic diagram of the extraction step 2004 and the tracking step 2006 of removing the region that is not the first region based on the flicker characteristic in the detection step 356 according to the embodiment of the present application. In step 2004, at least one respective first bright region is extracted from each binarized frame to obtain a location of each of the at least one respective first bright regions. In the example of fig. 22, in the binarized frame 2104, the regions 2112, 2114, and 2116 are the first bright regions after extraction. In the binarized frame 2202 after the binarized frame 2104, an area 2204 of the bright signal pulse corresponding to the first signal pulse of the invisible light signal transmitter 108 is the first bright area after extraction. In binarized frame 2206 following binarized frame 2202, area 2208 corresponding to the square area of the smart tv display is the first bright area after extraction. In the binarized frame 2210 after the binarized frame 2206, areas 2212 and 2214 corresponding to the bright signal pulse of the first signal pulse of the invisible light signal emitter 108 and the square area of the smart tv display are the first bright areas after extraction.

In step 2006, the location of each of the at least one respective first bright regions is tracked to detect each of at least one set of bright and dark regions corresponding to the binarized frame and being location-dependent, wherein the at least one set of bright and dark regions includes the first region. In the example of fig. 22, by way of tracking, the regions of the bright signal pulses (e.g., 2112, 2204, and 2212) corresponding to the first signal pulse of the invisible signal emitter 108 may be found to be correlated in position. Based on the locations of the regions 2112, 2204, and 2212, a first set of bright and dark regions corresponding to the binarized frame (e.g., 2104, 2202, 2206, and 2210) may be detected. One dark region of the first set of light and dark regions is represented by an area similar to the area (e.g., 2212) in, for example, binarized frame 2206. By way of tracking, areas (e.g., 2116, 2208, and 2214) corresponding to the square areas of the smart tv display may be found to be positionally related. Based on the locations of regions 2116, 2208, and 2214, a second set of bright and dark regions corresponding to the binarized frame (e.g., 2114, 2202, 2206, and 2210) may be detected. By way of tracking, it can be found that the region 2114 corresponding to the circular region of the smart tv display is positionally unrelated to any of the regions in each subsequent binarized frame (e.g., 2202, 2206, or 2210). Thus, based on the location of region 2114, a set of bright and dark regions corresponding to the binarized frame (e.g., 2114, 2202, 2206, and 2210) are not detected. Since the first and second sets of bright and dark regions satisfy the flicker characteristic, the first and second sets of bright and dark regions are candidate regions for identifying the first region (e.g., 2112, 2204, and 2212) of the data sequence. For region 2114, which corresponds to a circular region of the smart tv display, region 2114 will be disqualified as a candidate region for identifying the first region (e.g., 2112, 2204, and 2212) of the data sequence based on the nature of the flicker since no set of bright and dark regions can be detected.

Fig. 23 shows a flow chart of the marking and decoding step 358 in a method of analyzing a first frame to obtain a data sequence according to an embodiment of the application. Step 358 includes the following steps. Referring to fig. 23, in step 2302, at least one set of bright and dark regions is marked to obtain a corresponding second encoding pattern for each of the at least one set of bright and dark regions. In step 2304, a third encoding pattern of the corresponding second encoding pattern repeated in each of the at least one set of light and dark regions is determined that includes the first encoding pattern. In step 2306, the first encoding pattern is decoded to identify the data sequence.

FIG. 24 shows a schematic diagram of the marking 2302 and determination 2304 and decoding 2306 steps of removing regions that are not the first region based on repetition characteristics in the marking and decoding 358 step according to an embodiment of the application. In step 2302, at least one set of bright and dark regions is marked to obtain a corresponding second encoding pattern for each of the at least one set of bright and dark regions. The corresponding second coding pattern corresponds to a portion of the first frame having a predetermined duration, such as 1 second. In the example of FIG. 24, the second set of light and dark regions described with reference to FIG. 22 are marked to obtain a second encoding pattern 2402. The first set of light and dark regions described with reference to fig. 22 are marked to obtain the second encoding pattern 2404. In step 2304, a third encoding pattern of the corresponding second encoding pattern repeated in each of the at least one set of light and dark regions is determined that includes the first encoding pattern. In the example of FIG. 24, the second encoding pattern 2402 cannot be broken down into portions such that some of the portions are repeated. While the second encoding pattern 2404 may be broken down into portions such that some of the portions 2406 and 2408 are repeated. Thus, the second set of light and dark regions will be disqualified as candidate regions for identifying the first region of the data sequence. While leaving only the first set of light and dark regions as candidate regions for the first region to be used for identifying the data sequence. Thus, the portion 2406 or 2408 is a third encoding pattern that includes the first encoding pattern. In step 2306, the first encoding pattern is decoded to identify the data sequence. In the example of fig. 24, either portion 2406 or 2408 is decoded to identify data sequence 2410.

Referring to fig. 3, in step 360, a pairing request including the data sequence is transmitted to the transmitting device 100. In step 308, the pairing request is received from the receiving device 200. In step 310, a pairing request response is sent to the receiving device 200. In step 362, the pairing request response is received from the transmitting device 100. In step 364, the receiving device 200 pairs with the transmitting device based on the data sequence. In step 312, the transmitting device 200 pairs with the receiving device based on the data sequence. The communication method described with reference to fig. 3 to 24 is applicable not only to the case where the reception apparatus 200 is not paired with any transmission apparatus but also to the case where the reception apparatus 200 is paired with one transmission apparatus but is turned to be paired with the transmission apparatus 100, and is applicable to the case where the reception apparatus 200 has been paired with the transmission apparatus 100, is turned to be paired with another transmission apparatus, and is then turned back to be paired with the transmission apparatus 100.

Certain embodiments have one or a combination of the following features and/or advantages. The method includes successfully sampling using a signal pulse rate substantially the same as a camera frame rate, reducing an exposure time interval to improve a sampling failure condition and to improve visibility of a portion of a first region in a first frame, adjusting a white balance to make the portion of the first region more distinctive in color, automatically focusing on a first invisible light signal emitter to make it possible to focus on the portion of the first region, analyzing the first frame to remove a region that is not the first region based on brightness or color, flicker characteristics, and repeatability to identify a data sequence, and thus enabling invisible light communication using a visible light camera.

One of ordinary skill in the art will appreciate that each of the elements, modules, layers, blocks, algorithms, and steps of the systems or methods described and disclosed in the embodiments of the present application may be implemented using hardware, firmware, software, or combinations thereof. Whether such functionality is implemented as hardware, firmware, or software depends upon the application and design constraints imposed on the implementation. Skilled artisans may implement the functionality in varying ways for each particular application, and such implementation decisions should not be interpreted as causing a departure from the scope of the present application.

It should be understood that the systems and methods disclosed in the embodiments of the present application may be implemented in other ways. The embodiments described above are merely illustrative. The division of these modules is based solely on logical functions, and other divisions exist in implementations. These modules may or may not be physical modules. Multiple modules may be combined or integrated into one physical module. Any module may also be divided into a plurality of physical modules. Some features may also be omitted or skipped. On the other hand, the mutual coupling, direct coupling or communicative coupling shown or discussed is operated through some ports, devices or modules, whether indirectly or communicatively operated through electrical, mechanical or other types of means.

Modules used for purposes of explanation as discrete components may or may not be physically separated. These modules are located in one place or distributed over multiple network modules. Some or all of the modules are used for the purposes of these embodiments.

If the software functional module is implemented, used, and sold as a product, it may be stored in a computer-readable storage medium. Based on such understanding, the technical solutions proposed in the present application can be implemented in the form of software products in nature or in part. Alternatively, a part of the technical solution that is advantageous to the prior art may be implemented in the form of a software product. The software product is stored in a computer readable storage medium and includes a plurality of commands for at least one processor on the system to perform all or a portion of the steps disclosed in embodiments of the present application. The storage medium includes a USB disk, a removable hard disk, a Read Only Memory (ROM), a Random Access Memory (RAM), a floppy disk, or other medium capable of storing program code.

While the application has been described in connection with what is presently considered to be the most practical and preferred embodiment, it is to be understood that the application is not limited to the disclosed embodiment, but is intended to cover various arrangements made without departing from the scope of the broadest interpretation of the appended claims.

53页详细技术资料下载
上一篇:一种医用注射器针头装配设备
下一篇:对待通过光通信路径传输的数据进行基于功率的编码

网友询问留言

已有0条留言

还没有人留言评论。精彩留言会获得点赞!

精彩留言,会给你点赞!