课程咨询

雅思备考规划

扫码添加助教免费咨询雅思备考规划

扫码关注回复雅思获取最新雅思口语题库和备考资料

雅思阅读材料:Eye robot

2017-05-01 15:52:00来源:网络 柯林斯词典

  新东方在线雅思网为大家带来了雅思阅读材料:Eye robot,希望以下内容能够为同学们的雅思备考提供帮助。新东方在线雅思网将第一时间为大家发布最新、最全、最专业的雅思报名官网消息和雅思考试真题及解析,供大家参考。

  Poor eyesight remains one of the main obstacles to letting robots loose among humans. But it is improving, in part by aping natural vision

  ROBOTS are getting smarter and more agile all the time. They disarm bombs, fly combat missions, put together complicated machines, even play football. Why, then, one might ask, are they nowhere to be seen, beyond war zones, factories and technology fairs? One reason is that they themselves cannot see very well. And people are understandably wary of purblind contraptions bumping into them willy-nilly in the street or at home.

  All that a camera-equipped computer “sees” is lots of picture elements, or pixels. A pixel is merely a number reflecting how much light has hit a particular part of a sensor. The challenge has been to devise algorithms that can interpret such numbers as scenes composed of different objects in space. This comes naturally to people and, barring certain optical illusions, takes no time at all as well as precious little conscious effort. Yet emulating this feat in computers has proved tough.

  In natural vision, after an image is formed in the retina it is sent to an area at the back of the brain, called the visual cortex, for processing. The first nerve cells it passes through react only to simple stimuli, such as edges slanting at particular angles. They fire up other cells, further into the visual cortex, which react to simple combinations of edges, such as corners. Cells in each subsequent area discern ever more complex features, with those at the top of the hierarchy responding to general categories like animals and faces, and to entire scenes comprising assorted objects. All this takes less than a tenth of a second.

  The outline of this process has been known for years and in the late 1980s Yann LeCun, now at New York University, pioneered an approach to computer vision that tries to mimic the hierarchical way the visual cortex is wired. He has been tweaking his “convolutional neural networks” (ConvNets) ever since.

  Seeing is believing

  A ConvNet begins by swiping a number of software filters, each several pixels across, over the image, pixel by pixel. Like the brain’s primary visual cortex, these filters look for simple features such as edges. The upshot is a set of feature maps, one for each filter, showing which patches of the original image contain the sought-after element. A series of transformations is then performed on each map in order to enhance it and improve the contrast. Next, the maps are swiped again, but this time rather than stopping at each pixel, the filter takes a snapshot every few pixels. That produces a new set of maps of lower resolution. These highlight the salient features while reining in computing power. The whole process is then repeated, with several hundred filters probing for more elaborate shapes rather than just a few scouring for simple ones. The resulting array of feature maps is run through one final set of filters. These classify objects into general categories, such as pedestrians or cars.

  Many state-of-the-art computer-vision systems work along similar lines. The uniqueness of ConvNets lies in where they get their filters. Traditionally, these were simply plugged in one by one, in a laborious manual process that required an expert human eye to tell the machine what features to look for, in future, at each level. That made systems which relied on them good at spotting narrow classes of objects but inept at discerning anything else.

  Dr LeCun’s artificial visual cortex, by contrast, lights on the appropriate filters automatically as it is taught to distinguish the different types of object. When an image is fed into the unprimed system and processed, the chances are it will not, at first, be assigned to the right category. But, shown the correct answer, the system can work its way back, modifying its own parameters so that the next time it sees a similar image it will respond appropriately. After enough trial runs, typically 10,000 or more, it makes a decent fist of recognising that class of objects in unlabelled images.

  This still requires human input, though. The next stage is “unsupervised” learning, in which instruction is entirely absent. Instead, the system is shown lots of pictures without being told what they depict. It knows it is on to a promising filter when the output image resembles the input. In a computing sense, resemblance is gauged by the extent to which the input image can be recreated from the lower-resolution output. When it can, the filters the system had used to get there are retained.

  In a tribute to nature’s nous, the lowest-level filters arrived at in this unaided process are edge-seeking ones, just as in the brain. The top-level filters are sensitive to all manner of complex shapes. Caltech-101, a database routinely used for vision research, consists of some 10,000 standardised images of 101 types of just such complex shapes, including faces, cars and watches. When a ConvNet with unsupervised pre-training is shown the images from this database it can learn to recognise the categories more than 70% of the time. This is just below what top-scoring hand-engineered systems are capable of—and those tend to be much slower.

  This approach (which Geoffrey Hinton of the University of Toronto, a doyen of the field, has dubbed “deep learning”) need not be confined to computer-vision. In theory, it ought to work for any hierarchical system: language processing, for example. In that case individual sounds would be low-level features akin to edges, whereas the meanings of conversations would correspond to elaborate scenes.

  For now, though, ConvNet has proved its mettle in the visual domain. Google has been using it to blot out faces and licence plates in its Streetview application. It has also come to the attention of DARPA, the research arm of America’s Defence Department. This agency provided Dr LeCun and his team with a small roving robot which, equipped with their system, learned to detect large obstacles from afar and correct its path accordingly—a problem that lesser machines often, as it were, trip over. The scooter-sized robot was also rather good at not running into the researchers. In a selfless act of scientific bravery, they strode confidently in front of it as it rode towards them at a brisk walking pace, only to see it stop in its tracks and reverse. Such machines may not quite yet be ready to walk the streets alongside people, but the day they can is surely not far off.


本文关键字: 雅思阅读

为你特别匹配的雅思超值课程,祝你和雅思分手!
  • 新东方5月雅思公开讲座

    新东方雅思5月公开讲座

    新东方教师直播教你全科技巧!

    每天1小时

    查看详情
  • 雅思机考实战

    雅思机考实战

    剑桥雅思正版题目机考实战!

    每天1小时

    查看详情
  • 【知心雅思】6分录播课 (A类)

    【知心雅思】6分录播课 (A类)

    适合人群:想要冲6分的考生

    课时:434

    查看详情
  • 【知心雅思】6.5分录播课 (A类)

    【知心雅思】6.5分录播课 (A类)

    适合人群:想要冲6.5分的考生

    课时:464

    查看详情
  • 【知心雅思】7分录播课 (A类)

    【知心雅思】7分录播课 (A类)

    适合人群:想要冲7分的考生

    课时:443

    查看详情
雅思备考资料包

扫码添加助教

免费获取雅思备考资料包

更多资料
更多>>
  • 雅思备考英文书推荐汇总

    很多考鸭在备考过程中会想去寻找一些原文资料,想要轻松有效地提升自己的雅思水平,这是非常不错的做法! 本文主要为大家介绍实用的英文书籍,希望对大家的雅思备考有所帮助。

    来源 : 网络综合整理 2023-12-06 22:26:00 关键字 : 雅思备考 英文书 英文原籍

  • 雅思备考英文书推荐(六)

    很多考鸭在备考过程中会想去寻找一些原文资料,想要轻松有效地提升自己的雅思水平,这是非常不错的做法! 本文主要为大家介绍实用的英文书籍,希望对大家的雅思备考有所帮助。

    来源 : 网络综合整理 2023-12-06 22:23:00 关键字 : 雅思备考 英文书 英文原籍

  • 雅思备考英文书推荐(五)

    很多考鸭在备考过程中会想去寻找一些原文资料,想要轻松有效地提升自己的雅思水平,这是非常不错的做法! 本文主要为大家介绍实用的英文书籍,希望对大家的雅思备考有所帮助。

    来源 : 网络综合整理 2023-12-06 22:06:00 关键字 : 雅思备考 英文书 英文原籍

  • 雅思备考英文书推荐(四)

    很多考鸭在备考过程中会想去寻找一些原文资料,想要轻松有效地提升自己的雅思水平,这是非常不错的做法! 本文主要为大家介绍实用的英文书籍,希望对大家的雅思备考有所帮助。

    来源 : 网络综合整理 2023-12-06 21:36:00 关键字 : 雅思备考 英文书 英文原籍

  • 雅思备考英文书推荐(三)

    很多考鸭在备考过程中会想去寻找一些原文资料,想要轻松有效地提升自己的雅思水平,这是非常不错的做法! 本文主要为大家介绍实用的英文书籍,希望对大家的雅思备考有所帮助。

    来源 : 网络综合整理 2023-12-06 21:27:00 关键字 : 雅思备考 英文书 英文原籍

更多内容

移动学习

二维码

2024年1月-4月雅思口语题库

扫码添加助教号 回复【新题】 即可领取
更多>>
更多公开讲座>>

2024年雅思考试重点题

微信添加助教 回复【考试重点题】

助教微信
更多>>
更多资料