Geoffrey Hinton

Geoffrey Hinton at UBC.jpg
Hinton in 2013
Born
Geoffrey Everest Hinton

(1947-12-06) 6 December 1947 (age 75)[1]
Education
Known for
Awards
Scientific career
Fields
Institutions
ThesisRelaxation and its role in vision (1977)
Doctoral advisorChristopher Longuet-Higgins[3][4][5]
Doctoral students
Other notable students
Websitewww.cs.toronto.edu/~hinton/ Edit this at Wikidata

Geoffrey Everest Hinton CC FRS FRSC[12] (born 6 December 1947) is a British-Canadian cognitive psychologist and computer scientist, most noted for his work on artificial neural networks. Since 2013, he has divided his time working for Google (Google Brain) and the University of Toronto. In 2017, he co-founded and became the Chief Scientific Advisor of the Vector Institute in Toronto.[13][14]

With David Rumelhart and Ronald J. Williams, Hinton was co-author of a highly cited paper published in 1986 that popularised the backpropagation algorithm for training multi-layer neural networks,[15] although they were not the first to propose the approach.[16] Hinton is viewed as a leading figure in the deep learning community.[17][18][19][20][21] The dramatic image-recognition milestone of the AlexNet designed in collaboration with his students Alex Krizhevsky[22] and Ilya Sutskever for the ImageNet challenge 2012[23] was a breakthrough in the field of computer vision.[24]

Hinton received the 2018 Turing Award, together with Yoshua Bengio and Yann LeCun, for their work on deep learning.[25] They are sometimes referred to as the "Godfathers of AI" and "Godfathers of Deep Learning",[26][27] and have continued to give public talks together.[28][29]

Education

Hinton was educated at King's College, Cambridge, graduating in 1970 with a Bachelor of Arts in experimental psychology.[1] He continued his study at the University of Edinburgh where he was awarded a PhD in artificial intelligence in 1978 for research supervised by Christopher Longuet-Higgins.[3][30]

Career and research

After his PhD, he worked at the University of Sussex and, (after difficulty finding funding in Britain),[31] the University of California, San Diego and Carnegie Mellon University.[1] He was the founding director of the Gatsby Charitable Foundation Computational Neuroscience Unit at University College London[1] and is currently[32] a professor in the computer science department at the University of Toronto. He holds a Canada Research Chair in Machine Learning and is currently[when?] an advisor for the Learning in Machines & Brains program at the Canadian Institute for Advanced Research. Hinton taught a free online course on Neural Networks on the education platform Coursera in 2012.[33] Hinton joined Google in March 2013 when his company, DNNresearch Inc., was acquired. He is planning to "divide his time between his university research and his work at Google".[34]

Hinton's research investigates ways of using neural networks for machine learning, memory, perception and symbol processing. He has authored or co-authored over 200 peer reviewed publications.[2][35] At the Conference on Neural Information Processing Systems (NeuRIPS) 2022, Hinton introduced a new learning algorithm for neural networks that he calls the "Forward-Forward" algorithm. The idea of the new algorithm is to replace the traditional forward-backward passes of backpropagation with two forward passes, one with positive (i.e. real) data and the other with negative data which could be generated by the network itself.[36]

While Hinton was a professor at Carnegie Mellon University (1982–1987), David E. Rumelhart and Hinton and Ronald J. Williams applied the backpropagation algorithm to multi-layer neural networks. Their experiments showed that such networks can learn useful internal representations of data.[15] In an interview of 2018,[37] Hinton said that "David E. Rumelhart came up with the basic idea of backpropagation, so it's his invention." Although this work was important in popularising backpropagation, it was not the first to suggest the approach.[16] Reverse-mode automatic differentiation, of which backpropagation is a special case, was proposed by Seppo Linnainmaa in 1970, and Paul Werbos proposed to use it to train neural networks in 1974.[16]

During the same period, Hinton co-invented Boltzmann machines with David Ackley and Terry Sejnowski.[38] His other contributions to neural network research include distributed representations, time delay neural network, mixtures of experts, Helmholtz machines and Product of Experts. In 2007 Hinton coauthored an unsupervised learning paper titled Unsupervised learning of image transformations.[39] An accessible introduction to Geoffrey Hinton's research can be found in his articles in Scientific American in September 1992 and October 1993.[40]

In October and November 2017 respectively, Hinton published two open access research papers[41][42] on the theme of capsule neural networks, which according to Hinton are "finally something that works well."[43]

Notable former PhD students and postdoctoral researchers from his group include Peter Dayan,[44] Sam Roweis,[44] Max Welling,[44] Richard Zemel,[3][6] Brendan Frey,[7] Radford M. Neal,[8] Yee Whye Teh,[9] Ruslan Salakhutdinov,[10] Ilya Sutskever,[11] Yann LeCun,[45] Alex Graves,[44] and Zoubin Ghahramani.

Honours and awards

From left to right Russ Salakhutdinov, Richard S. Sutton, Geoffrey Hinton, Yoshua Bengio and Steve Jurvetson in 2016
From left to right Russ Salakhutdinov, Richard S. Sutton, Geoffrey Hinton, Yoshua Bengio and Steve Jurvetson in 2016

Hinton was elected a Fellow of the Royal Society (FRS) in 1998.[12] He was the first winner of the Rumelhart Prize in 2001.[46] His certificate of election for the Royal Society reads:

Geoffrey E. Hinton is internationally distinguished for his work on artificial neural nets, especially how they can be designed to learn without the aid of a human teacher. This may well be the start of autonomous intelligent brain-like machines. He has compared effects of brain damage with effects of losses in such a net, and found striking similarities with human impairment, such as for recognition of names and losses of categorisation. His work includes studies of mental imagery, and inventing puzzles for testing originality and creative intelligence. It is conceptual, mathematically sophisticated and experimental. He brings these skills together with striking effect to produce important work of great interest.[47]

In 2001, Hinton was awarded an honorary doctorate from the University of Edinburgh.[48] He was the 2005 recipient of the IJCAI Award for Research Excellence lifetime-achievement award.[49] He has also been awarded the 2011 Herzberg Canada Gold Medal for Science and Engineering.[50] In 2013, Hinton was awarded an honorary doctorate from the Université de Sherbrooke.[51]

In 2016, he was elected a foreign member of National Academy of Engineering "For contributions to the theory and practice of artificial neural networks and their application to speech recognition and computer vision".[52] He also received the 2016 IEEE/RSE Wolfson James Clerk Maxwell Award.[53]

He has won the BBVA Foundation Frontiers of Knowledge Award (2016) in the Information and Communication Technologies category "for his pioneering and highly influential work" to endow machines with the ability to learn.[54]

Together with Yann LeCun, and Yoshua Bengio, Hinton won the 2018 Turing Award for conceptual and engineering breakthroughs that have made deep neural networks a critical component of computing.[55][56][57]

In 2018, he was awarded a Companion of the Order of Canada.[58] In 2022 he received the Princess of Asturias Award in the category "Scientific Research", along with Yann LeCun, Yoshua Bengio, and Demis Hassabis.[59]

Personal life

Hinton is the great-great-grandson of the mathematician and educator Mary Everest Boole and her husband, the logician George Boole,[60] whose work eventually became one of the foundations of modern computer science. Another great-great-grandfather was the surgeon and author James Hinton,[61] who was the father of Charles Howard Hinton. Hinton's father was Howard Hinton.[1][62] His middle name comes from another relative, George Everest.[31] He is the nephew of the economist Colin Clark.[63] He lost his second wife to ovarian cancer in 1994.[64]

Views

Hinton moved from the U.S. to Canada in part due to disillusionment with Ronald Reagan-era politics and disapproval of military funding of artificial intelligence.[31]

Hinton has petitioned against lethal autonomous weapons. Regarding existential risk from artificial intelligence, Hinton typically declines to make predictions more than five years into the future, noting that exponential progress makes the uncertainty too great.[65]

Hinton is optimistic about AI’s impact on the job market: “The phrase ‘artificial general intelligence’ carries with it the implication that this sort of single robot is suddenly going to be smarter than you. I don’t think it’s going to be that. I think more and more of the routine things we do are going to be replaced by AI systems — like the Google Assistant.” [66]

Hinton argues that AGI won't make humans redundant. Rather, he says, it will remain for the most part myopic in its understanding of the world – at least in the near future. He believes that it’ll continue to improve our lives in small but meaningful ways. "[AI in the future is] going to know a lot about what you’re probably going to want to do and how to do it, and it’s going to be very helpful. But it’s not going to replace you," he said. "If you took [a] system that was developed to be able to be very good [at driving], and you sent it on its first date, I think it would be a disaster." And for dangerous tasks currently performed by humans, that's a step in the right direction, according to Hinton.[67]

In an interview[68] with CBS on March 25, 2023, Hinton made the following comment about the recent progress in AI: "I think it’s comparable in scale with the Industrial Revolution or electricity — or maybe the wheel."[69]

Other view expressed in the CBS interview[68]
Notable Statement/Quote Impact
"We're making machines that are going to be superhuman at a wide range of things." Hinton believes that AI technology will surpass human abilities in many areas, which could have significant implications for society.
"AI is going to change the world more than anything in the history of humanity. More than electricity." Hinton sees AI as a transformative technology that could rival the impact of the Industrial Revolution and other major historical events.
"The kind of intelligence we're developing is very different from our intelligence. So it's this idiot-savant kind of intelligence." Hinton acknowledges that AI has a very different kind of intelligence from humans, which may limit its ability to fully understand human experiences and emotions.
"I think whatever is going to happen is pretty much inevitable... One person stopping doing research wouldn't stop this happening." Hinton believes that the development of AI is inevitable, and that researchers must think about how to control it responsibly.
"I think it's going to make jobs different. People are going to be doing the more creative end. Less of the routine end." Hinton predicts that AI will change the nature of many jobs, with humans focusing more on creative tasks and less on routine tasks.
"We've seen that as they scale up chat GPT. It's not radically new ideas there, it's just more connections and more data to train it with." Hinton notes that some recent progress in AI has come from simply scaling up existing models with more data and more computing power, rather than fundamentally new ideas.
"Time to prepare would be good. And so I think it's very reasonable for people to be worrying about those issues now, even though it's not going to happen in the next year or two." Hinton believes that people should be thinking about the potential negative impacts of AI, such as job displacement and ethical concerns, even though they may not be fully realized for several years.

References

  1. ^ a b c d e Anon (2015) "Hinton, Prof. Geoffrey Everest". Who's Who. ukwhoswho.com (online Oxford University Press ed.). A & C Black, an imprint of Bloomsbury Publishing plc. (Subscription or UK public library membership required.) doi:10.1093/ww/9780199540884.013.20261
  2. ^ a b Geoffrey Hinton publications indexed by Google Scholar Edit this at Wikidata
  3. ^ a b c Geoffrey Hinton at the Mathematics Genealogy Project
  4. ^ Geoffrey E. Hinton's Academic Genealogy
  5. ^ Gregory, R. L.; Murrell, J. N. (2006). "Hugh Christopher Longuet-Higgins. 11 April 1923 -- 27 March 2004: Elected FRS 1958". Biographical Memoirs of Fellows of the Royal Society. 52: 149–166. doi:10.1098/rsbm.2006.0012.
  6. ^ a b Zemel, Richard Stanley (1994). A minimum description length framework for unsupervised learning (PhD thesis). University of Toronto. OCLC 222081343. ProQuest 304161918.
  7. ^ a b Frey, Brendan John (1998). Bayesian networks for pattern classification, data compression, and channel coding (PhD thesis). University of Toronto. OCLC 46557340. ProQuest 304396112.
  8. ^ a b Neal, Radford (1995). Bayesian learning for neural networks (PhD thesis). University of Toronto. OCLC 46499792. ProQuest 304260778.
  9. ^ a b Whye Teh, Yee (2003). Bethe free energy and contrastive divergence approximations for undirected graphical models. utoronto.ca (PhD thesis). University of Toronto. hdl:1807/122253. OCLC 56683361. ProQuest 305242430.
  10. ^ a b Salakhutdinov, Ruslan (2009). Learning deep generative models (PhD thesis). University of Toronto. ISBN 9780494610800. OCLC 785764071. ProQuest 577365583.
  11. ^ a b Sutskever, Ilya (2013). Training Recurrent Neural Networks. utoronto.ca (PhD thesis). University of Toronto. hdl:1807/36012. OCLC 889910425. ProQuest 1501655550.
  12. ^ a b Anon (1998). "Professor Geoffrey Hinton FRS". London: Royal Society. Archived from the original on 3 November 2015. One or more of the preceding sentences incorporates text from the royalsociety.org website where:

    "All text published under the heading 'Biography' on Fellow profile pages is available under Creative Commons Attribution 4.0 International License." --"Royal Society Terms, conditions and policies". Archived from the original on 11 November 2016. Retrieved 9 March 2016.((cite web)): CS1 maint: bot: original URL status unknown (link)

  13. ^ Hernandez, Daniela (7 May 2013). "The Man Behind the Google Brain: Andrew Ng and the Quest for the New AI". Wired. Retrieved 10 May 2013.
  14. ^ "Geoffrey E. Hinton – Google AI". Google AI.
  15. ^ a b Rumelhart, David E.; Hinton, Geoffrey E.; Williams, Ronald J. (9 October 1986). "Learning representations by back-propagating errors". Nature. 323 (6088): 533–536. Bibcode:1986Natur.323..533R. doi:10.1038/323533a0. ISSN 1476-4687. S2CID 205001834.
  16. ^ a b c Schmidhuber, Jürgen (1 January 2015). "Deep learning in neural networks: An overview". Neural Networks. 61: 85–117. arXiv:1404.7828. doi:10.1016/j.neunet.2014.09.003. PMID 25462637. S2CID 11715509.
  17. ^ "Geoffrey Hinton was briefly a Google intern in 2012 because of bureaucracy – TechCrunch". techcrunch.com. 14 September 2017. Retrieved 28 March 2018.
  18. ^ Somers, James. "Progress in AI seems like it's accelerating, but here's why it could be plateauing". MIT Technology Review. Retrieved 28 March 2018.
  19. ^ "How U of T's 'godfather' of deep learning is reimagining AI". University of Toronto News. Retrieved 28 March 2018.
  20. ^ "'Godfather' of deep learning is reimagining AI". Retrieved 28 March 2018.
  21. ^ "Geoffrey Hinton, the 'godfather' of deep learning, on AlphaGo". Maclean's. 18 March 2016. Retrieved 28 March 2018.
  22. ^ Gershgorn, Dave (18 June 2018). "The inside story of how AI got good enough to dominate Silicon Valley". Quartz. Retrieved 5 October 2018.
  23. ^ Krizhevsky, Alex; Sutskever, Ilya; Hinton, Geoffrey E. (3 December 2012). "ImageNet classification with deep convolutional neural networks". Nips'12. Curran Associates Inc.: 1097–1105. ((cite journal)): Cite journal requires |journal= (help)
  24. ^ "How a Toronto professor's research revolutionized artificial intelligence | Toronto Star". Toronto Star. Retrieved 13 March 2018.
  25. ^ 27 Mar, Emily Chung · CBC News · Posted; March 27, 2019 6:00 am ET | Last Updated. "Canadian researchers who taught AI to learn like humans win $1M award | CBC News". Canadian Broadcasting Corporation. Retrieved 27 March 2019.
  26. ^ Ranosa, Ted (29 March 2019). "Godfathers Of AI Win This Year's Turing Award And $1 Million". Tech Times. Retrieved 5 November 2020.
  27. ^ Shead, Sam. "The 3 'Godfathers' Of AI Have Won The Prestigious $1M Turing Prize". Forbes. Retrieved 5 November 2020.
  28. ^ Ray, Tiernan. "Nvidia's GTC will feature deep learning cabal of LeCun, Hinton, and Bengio". ZDNet. Retrieved 7 April 2021.
  29. ^ "50 Years at CMU: The Inaugural Raj Reddy Artificial Intelligence Lecture". cs.cmu.edu. Retrieved 2 March 2022.
  30. ^ Hinton, Geoffrey Everest (1977). Relaxation and its role in vision. ed.ac.uk (PhD thesis). University of Edinburgh. hdl:1842/8121. OCLC 18656113. EThOS uk.bl.ethos.482889. icon of an open green padlock
  31. ^ a b c Smith, Craig S. (23 June 2017). "The Man Who Helped Turn Toronto into a High-Tech Hotbed". The New York Times. Retrieved 27 June 2017.
  32. ^ https://www.cs.toronto.edu/~hinton/fullcv.pdf[bare URL PDF]
  33. ^ "Archived copy". Archived from the original on 31 December 2016. Retrieved 30 December 2016.((cite web)): CS1 maint: archived copy as title (link)
  34. ^ "U of T neural networks start-up acquired by Google" (Press release). Toronto, ON. 12 March 2013. Retrieved 13 March 2013.
  35. ^ Geoffrey Hinton publications indexed by the Scopus bibliographic database. (subscription required)
  36. ^ "The Forward-Forward Algorithm: Some Preliminary Investigations" (PDF) (Press release). Toronto, ON.
  37. ^ Ford, Martin (2018). Architects of Intelligence: The truth about AI from the people building it. Packt Publishing. ISBN 978-1-78913-151-2.
  38. ^ Ackley, David H; Hinton Geoffrey E; Sejnowski, Terrence J (1985), "A learning algorithm for Boltzmann machines", Cognitive science, Elsevier, 9 (1): 147–169
  39. ^ Hinton, Geoffrey E. "Geoffrey E. Hinton's Publications in Reverse Chronological Order".
  40. ^ "Stories by Geoffrey E. Hinton in Scientific American". Scientific American.
  41. ^ Sabour, Sara; Frosst, Nicholas; Hinton, Geoffrey. October 2017. "Dynamic Routing Between Capsules"
  42. ^ "Matrix capsules with EM routing" 3 November 2017. OpenReview.net
  43. ^ Geib, Claudia. 2 November 2017. "We’ve Finally Created an AI Network That’s Been Decades in the Making" Futurism.com
  44. ^ a b c d "Geoffrey Hinton's postdocs". Geoffrey Hinton.
  45. ^ "Yann LeCun's Research and Contributions". yann.lecun.com. Retrieved 13 March 2018.
  46. ^ "Current and Previous Recipients". David E. Rumelhart Prize. Archived from the original on 2 March 2017.
  47. ^ Anon (1998). "Certificate of election EC/1998/21: Geoffrey Everest Hinton". London: Royal Society. Archived from the original on 5 May 2017.
  48. ^ "Distinguished Edinburgh graduate receives ACM A.M. Turing Award". The University of Edinburgh. Retrieved 9 April 2019.
  49. ^ "IJCAI Awards | IJCAI". www.ijcai.org. Retrieved 5 August 2021.
  50. ^ "Artificial intelligence scientist gets M prize". CBC News. 14 February 2011.
  51. ^ "Geoffrey Hinton, keystone researcher in artificial intelligence, visits the Université de Sherbrooke". Université de Sherbrooke. 19 February 2014.
  52. ^ "National Academy of Engineering Elects 80 Members and 22 Foreign Members". NAE. 8 February 2016.
  53. ^ "2016 IEEE Medals and Recognitions Recipients and Citations" (PDF). IEEE. Retrieved 7 July 2016.
  54. ^ "The BBVA Foundation bestows its award on the architect of the first machines capable of learning the way people do". BBVA Foundation. 17 January 2017.
  55. ^ "Vector Institutes Chief Scientific Advisor Dr.Geoffrey Hinton Receives ACM A.M. Turing Award Alongside Dr.Yoshua Bengio and Dr.Yann Lecun". NAE. 27 March 2019.
  56. ^ "Three Pioneers in Artificial Intelligence Win Turing Award". The New York Times. 27 March 2019. Retrieved 27 March 2019.
  57. ^ "Fathers of the Deep Learning Revolution Receive ACM A.M. Turing Award - Bengio, Hinton and LeCun Ushered in Major Breakthroughs in Artificial Intelligence". Association for Computing Machinery. 27 March 2019. Retrieved 27 March 2019.
  58. ^ "Governor General Announces 103 New Appointments to the Order of Canada, December 2018". 21 December 2018. Archived from the original on 19 November 2019. Retrieved 7 June 2020.
  59. ^ Princess of Asturias Awards 2022
  60. ^ "Geoffrey Hinton: The story of the British 'Godfather of AI' - who's not sat down since 2005". Sky News. Retrieved 7 April 2021.
  61. ^ The Isaac Newton of logic
  62. ^ Salt, George (1978). "Howard Everest Hinton. 24 August 1912-2 August 1977". Biographical Memoirs of Fellows of the Royal Society. 24: 150–182. doi:10.1098/rsbm.1978.0006. ISSN 0080-4606.
  63. ^ Shute, Joe (26 August 2017). "The 'Godfather of AI' on making machines clever and whether robots really will learn to kill us all?". The Daily Telegraph. Retrieved 20 December 2017.
  64. ^ Shute, Joe (26 August 2017). "The 'Godfather of AI' on making machines clever and whether robots really will learn to kill us all?". The Telegraph. Retrieved 30 January 2018.
  65. ^ name="fogofprogress">Hinton, Geoffrey. "Lecture 16d The fog of progress" (PDF).
  66. ^ "Geoffrey Hinton and Demis Hassabis: AGI is nowhere close to being a reality". venturebeat. 17 December 2018.
  67. ^ "Geoffrey Hinton and Demis Hassabis: AGI is nowhere close to being a reality". venturebeat. 17 December 2018.
  68. ^ a b ""Godfather of artificial intelligence" talks impact and potential of new AI". CBS News. CBS. 25 March 2023. Retrieved 28 March 2023.
  69. ^ "Geoffrey Hinton, alumnus Nick Frosst discuss impact and potential of new AI on CBS News". Department of Computer Science, University of Toronto. Retrieved 28 March 2023.