the table summarizing the eight authors of “Attention Is All You Need” (2017), including links to key papers and their current achievements:
Name | Main Papers / Research (Links) | Current Achievements / Activities | Notes |
---|---|---|---|
Ashish Vaswani | “Attention Is All You Need” (2017) → arXiv PDF “Bottleneck Transformers for Visual Recognition” → DeepAI Profile | Co-founder and CEO of Essential AI. Former Google Brain researcher. Now focused on efficient AI infrastructure and social impact. (Wikipedia) | One of the lead designers of the Transformer architecture. |
Noam Shazeer | “Attention Is All You Need” (2017) → arXiv “Switch Transformer: Scaling to Trillion Parameter Models” (2022) → JMLR Paper | Technical co-lead for Google’s Gemini AI project. Previously founded Character.AI before returning to Google. (Softwarereport) | Deeply involved in multi-head self-attention and large-scale model design. |
Niki Parmar | “Attention Is All You Need” (2017) → arXiv | Research scientist at Google and co-founder of Essential AI. (Communications Today) | Transitioned from research to entrepreneurship in AI. |
Jakob Uszkoreit | “Attention Is All You Need” (2017) → arXiv | Co-founder of biotech startup Inceptive, working on AI for molecular biology. (USC Viterbi School News) | Moved from machine translation research into bio-AI applications. |
Llion Jones | “Attention Is All You Need” (2017) → arXiv | CTO of Sakana AI, a Tokyo-based startup. Former Google Research scientist. (Hulkapps Profile) | Active in AI research and entrepreneurship in Asia. |
Aidan N. Gomez | “Attention Is All You Need” (2017) → arXiv | CEO and co-founder of Cohere, a Toronto-based AI company developing language models for enterprises. (New Yorker Feature) | A prominent example of a Transformer author turned successful AI entrepreneur. |
Łukasz Kaiser | “Attention Is All You Need” (2017) → arXiv “Fast Decoding in Sequence Models using Discrete Latent Variables” (2018) → arXiv Link | Research scientist at OpenAI, formerly at Google Brain. (Wired Feature) | Focuses on core model architecture and AI efficiency research. |
Illia Polosukhin | “Attention Is All You Need” (2017) → arXiv | Co-founder of NEAR Protocol, a blockchain and Web3 infrastructure project. (Wired Profile) | Former Google AI researcher; now bridging AI and decentralized technologies. |
Comments
Post a Comment