Back to news

June 14, 2022 | International, Clean technologies, Big data and Artifical Intelligence, Advanced manufacturing 4.0, Autonomous systems (Drones / E-VTOL), Virtual design and testing, Additive manufacturing

Startup Genome

Startup ecosystem development through data. We advise world leaders in policymaking, strategy and actions to drive innovation and economic growth.

Key Insights from #GSER2022

  • The same five ecosystems remain at the top of the ranking as in 2020 and 2021, but Beijing has dropped one place, with Boston taking its former place at #4. Silicon Valley is #1, followed by New York City and London tied at #2, Boston at #4, and Beijing at #5.
  • Seoul entered the global top 10 ecosystems for the first time, up six places from #16 in 2021 and #20 in 2020.
  • Several Indian ecosystems have risen in the rankings, most notably Delhi, which is 11 places higher than in 2021, entering the top 30 for the first time at #26. Bangalore has moved up one place from last year, to #22.
  • Overall, China's ecosystems have declined in the rankings, a reflection of the relative decline in early-stage funding in comparison to other ecosystems.
  • Helsinki has risen more than 20 places from last year, joining the runners-up category at joint #35.
  • A record 540 companies achieved unicorn status in 2021, up from 150 in 2020.
  • In 2021, Brazil saw 237% growth in the dollar amount of Series B+ rounds compared to 2020. The nation's total exit amount for 2021 was $49 billion, a huge leap from $1 billion in 2020.
  • Asia experienced a 312% increase in the dollar amount of exits over $50 million from 2020 to 2021.
  • In 2021, the dollar amount of exits in London grew 413% from 2020. The ecosystem's Series B+ rounds increased 162% in terms of dollar amount from 2020, and it saw 55% more $50 million+ exits in 2021 versus 2020.

https://startupgenome.com/report/gser2022

On the same subject

  • Trustworthy AI: A Conversation with NIST's Chuck Romine

    January 21, 2020

    Trustworthy AI: A Conversation with NIST's Chuck Romine

    By: Charles Romine Artificial Intelligence (AI) promises to grow the economy and improve our lives, but with these benefits, it also brings new risks that society is grappling with. How can we be sure this new technology is not just innovative and helpful, but also trustworthy, unbiased, and resilient in the face of attack? We sat down with NIST Information Technology Lab Director Chuck Romine to learn how measurement science can help provide answers. How would you define artificial intelligence? How is it different from regular computing? One of the challenges with defining artificial intelligence is that if you put 10 people in a room, you get 11 different definitions. It's a moving target. We haven't converged yet on exactly what the definition is, but I think NIST can play an important role here. What we can't do, and what we never do, is go off in a room and think deep thoughts and say we have the definition. We engage the community. That said, we're using a narrow working definition specifically for the satisfaction of the Executive Order on Maintaining American Leadership in Artificial Intelligence, which makes us responsible for providing guidance to the federal government on how it should engage in the standards arena for AI. We acknowledge that there are multiple definitions out there, but from our perspective, an AI system is one that exhibits reasoning and performs some sort of automated decision-making without the interference of a human. There's a lot of talk at NIST about “trustworthy” AI. What is trustworthy AI? Why do we need AI systems to be trustworthy? AI systems will need to exhibit characteristics like resilience, security and privacy if they're going to be useful and people can adopt them without fear. That's what we mean by trustworthy. Our aim is to help ensure these desirable characteristics. We want systems that are capable of either combating cybersecurity attacks, or, perhaps more importantly, at least recognizing when they are being attacked. We need to protect people's privacy. If systems are going to operate in life-or-death type of environments, whether it's in medicine or transportation, people need to be able to trust AI will make the right decisions and not jeopardize their health or well-being. Resilience is important. An artificial intelligence system needs to be able to fail gracefully. For example, let's say you train an artificial intelligence system to operate in a certain environment. Well, what if the system is taken out of its comfort zone, so to speak? One very real possibility is catastrophic failure. That's clearly not desirable, especially if you have the AI deployed in systems that operate critical infrastructure or our transportation systems. So, if the AI is outside of the boundaries of its nominal operating environment, can it fail in such a way that it doesn't cause a disaster, and can it recover from that in a way that allows it to continue to operate? These are the characteristics that we're looking for in a trustworthy artificial intelligence system. NIST is supposed to be helping industry before they even know they needed us to. What are we thinking about in this area that is beyond the present state of development of AI? Industry has a remarkable ability to innovate and to provide new capabilities that people don't even realize that they need or want. And they're doing that now in the AI consumer space. What they don't often do is to combine that push to market with deep thought about how to measure characteristics that are going to be important in the future. And we're talking about, again, privacy, security and resilience ... trustworthiness. Those things are critically important, but many companies that are developing and marketing new AI capabilities and products may not have taken those characteristics into consideration. Ultimately, I think there's a risk of a consumer backlash where people may start saying these things are too easy to compromise and they're betraying too much of my personal information, so get them out of my house. What we can do to help, and the reason that we've prioritized trustworthy AI, is we can provide that foundational work that people in the consumer space need to manage those risks overall. And I think that the drumbeat for that will get increasingly louder as AI systems begin to be marketed for more than entertainment. Especially at the point when they start to operate critical infrastructure, we're going to need a little more assurance. That's where NIST can come together with industry to think about those things, and we've already had some conversations with industry about what trustworthy AI means and how we can get there. I'm often asked, how is it even possible to influence a trillion-dollar, multitrillion-dollar industry on a budget of $150 million? And the answer is, if we were sitting in our offices doing our own work independent of industry, we would never be able to. But that's not what we do. We can work in partnership with industry, and we do that routinely. And they trust us, they're thrilled when we show up, and they're eager to work with us. AI is a scary idea for some people. They've seen “I, Robot,” or “The Matrix,” or “The Terminator.” What would you say to help them allay these fears? I think some of this has been overhyped. At the same time, I think it's important to acknowledge that risks are there, and that they can be pretty high if they're not managed ahead of time. For the foreseeable future, however, these systems are going to be too fragile and too dependent on us to worry about them taking over. I think the biggest revolution is not AI taking over, but AI augmenting human intelligence. We're seeing examples of that now, for instance, in the area of face recognition. The algorithms for face recognition have improved at an astonishing rate over the last seven years. We're now at the point where, under controlled circumstances, the best artificial intelligence algorithms perform on par with the best human face recognizers. A fascinating thing we learned recently, and published in a report, is that if you take two trained human face recognizers and put them together, the dual system doesn't perform appreciably better than either one of them alone. If you take two top-performing algorithms, the combination of the two doesn't really perform much better than either one of them alone. But if you put the best algorithm together with a trained recognizer, that system performs substantially better than either one of them alone. So, I think, human augmentation by AI is going to be the revolution. What's next? I think one of the things that is going to be necessary for us is pulling out the desirable characteristics like usability, interoperability, resilience, security, privacy and all the things that will require a certain amount of care to build into the systems, and get innovators to start incorporating them. Guidance and standards can help to do that. Last year, we published our plan for how the federal government should engage in the AI standards development process. I think there's general agreement that guidance will be needed for interoperability, security, reliability, robustness, these characteristics that we want AI systems to exhibit if they're going to be trusted. https://www.nist.gov/blogs/taking-measure/trustworthy-ai-conversation-nists-chuck-romine

  • BIND 4.0 | Call for proposal

    July 20, 2022

    BIND 4.0 | Call for proposal

    The BIND 4.0 International Open Call for Startups to apply for the chance to put innovation into action with paid contracts with industry leaders, is now active and accepting applications! Who do we Work With? Young companies worldwide developing solutions that can be applied to the Intelligent Industry, Clean Energy & Sustainability, Health Tech and/or Food Tech sectors for our 65+ Corporate Partners based in the Basque Country, Spain. The technologies of interest are wide ranging and not limited to: AI, Cybersecurity, Big Data Analytics, IOT/IIOT, Smart Energy, AR/VR, Machine or Artificial Vision, Collaborative Robotics, Additive Manufacturing, Biobased Solutions, Clean Tech, New Materials, NanoTech, Medical Devices, Smart Sensors, etc. Why BIND 4.0 Open Innovation & Acceleration Program? It is a government sponsored initiative with 6 years of proven track record accelerating over 160 International Startups and catalyzing countless business opportunities. 45K Average Contract Amount 65+ Big Name Reference Clients High Impact Acceleration, Mentoring & Support Zero Equity Taken & No Fee to Participate Please Share our International Startup Open Call. If you'd like us to prepare something targeted to your newsletter or startup portfolio please let us know. Easily share our latest social media posts and upcoming Webinar events via the links below. Key Details and links: Startup Open Call, Accepting Applications Deadline Sep 8, 2022 BIND 4.0 Webinar Presentation with Alumni July 28, 2022 BIND 4.0 Webinar Presentation with Alumni Aug 30, 2022 BIND 4.0 2022 Program Catalogue 2022 Catalogue BIND 4.0 Website For More Resources www.BIND40.com HTML Email Open Call Announcement Link to Email Message LinkedIn Open Call Post BIND 4.0 2022 Open Call Twitter Open Call Publication BIND 4.0 2022 Open Call We Value Collaborative Partnerships. Remember, for our Collaborative Partners with relevant topics and programs, we offer event promotion via our alumni newsletter or social media, targeted startup referrals, plus networking via invitations to innovation events. Have Any Startup Recommendations? Send us a quick email with any Startups that you think could be a good fit for our immersion, acceleration and open innovation program. Send their contact info so we can reach out to them directly and fast track their application based on your referral. Startups can set up a 1:1 Meeting with us, or if you'd rather send them this Link to Apply, the application deadline is September 8th.

  • New NATO Innovation Hub challenge: Trust in autonomous systems

    October 22, 2020

    New NATO Innovation Hub challenge: Trust in autonomous systems

    Hello, Similar to IDEaS, the NATO Innovation Hub is a community where experts from around the world collaborate to tackle NATO challenges and design solutions. The Hub has recently launched a challenge seeking innovative solutions that address how trust in autonomous systems can be established and strengthened. Solutions can include any combination of methodologies, concepts, techniques and technologies. These challenges are open to all, including the Canadian innovation community. Submit your solutions by November 17, 2020: https://www.innovationhub-act.org/challenge-intro If you have questions, contact the NATO Innovation Hub by email: contact@InnovationHub-act.org Thank you, The IDEaS Team

All news