25 janvier 2022 | International, Technologies propres

LANCEMENT DU PROJET INORTH

Le CRIAQ a récemment lancé le projet iNorth, financé par le Ministère de l'Économie et de l'Innovation et CRSNG, et dont l'objectif est de supporter le développement d'un nouveau système de surveillance et de contrôle de température de batteries au lithium-ion pour l'aérospatiale. Félicitations aux partenaires impliqués dans ce projet : CAE, Calogy Solutions, Université de Sherbrooke et ÉTS.

https://www.linkedin.com/feed/update/urn:li:activity:6889971693062774784/?mc_cid=0426e057c8&mc_eid=3dbf0b164e

Sur le même sujet

  • Blast, un programme soutenu par Starburst qui veut faire exploser les startups françaises du spatial

    27 novembre 2020

    Blast, un programme soutenu par Starburst qui veut faire exploser les startups françaises du spatial

    Plusieurs spécialistes de l'innovation se sont alliés pour créer Blast, le premier programme français d'accélération de jeunes pousses à fort contenu technologique dans les domaines de l'aéronautique, du spatial et de la défense afin de « faire émerger et d'accompagner une vingtaine de start-up par an ». Le programme réunit Starburst, premier accélérateur mondial dédié aux start-up de l'aéronautique, du spatial et de la défense (ASD), l'Office national d'études et de recherches aérospatiales (Onera), l'École Polytechnique et la SATT Paris-Saclay, dont le but est de développer la mise en oeuvre des innovations issues de la recherche académique. À l'origine du projet, qui sera formellement lancé en janvier, le constat du « manque d'un programme constitué en France dédié à l'accompagnement des projets deep tech (portant sur des technologies très avancées, NDLR) pour le secteur » de l'aéronautique, de la défense et du spatial, expliquent ses promoteurs dans un communiqué. Si les jeunes pousses dans ces domaines sont particulièrement actives aux États-Unis ou encore en Israël, le mouvement doit être « être renforcé en Europe et particulièrement en France », estiment-ils. D'autant que l'aéronautique, le spatial et la défense « portent par nature les caractéristiques du deep tech: de fortes barrières à l'entrée, un cycle de développement long et complexe et un caractère disruptif fort en cas de succès ». Des levées de fonds d'amorçage prévues Le programme doit permettre d'accompagner des projets correspondant aux besoins de l'industrie et de les pérenniser par des contrats commerciaux et des collaborations avec les industriels ou instituts publics. « L'idée, c'est de créer plus de start-up pour alimenter les futurs programmes aéronautiques, spatiaux et de défense », résume François Chopard, fondateur de Starburst, qui assurera la coordination du programme. Les domaines visés sont notamment ceux de l'aviation décarbonée, les plateformes de mobilité aérienne urbaines, l'intelligence artificielle, les technologies portant sur l'autonomie (capteurs, fusion de données), les services informatiques dématérialisés (cloud) sécurisés ou encore le spatial. Des levées de fonds « d'amorçage de deux à trois millions d'euros » sont prévues. Blast est l'un des dix programmes sélectionnés par le gouvernement à l'issue d'un appel à projets d'accompagnement de start-up à fort contenu technologique dans le cadre du Programme d'investissements d'avenir (PIA) Doté d'un montant maximal global de 9,6 millions d'euros, le soutien public, via Bpifrance, permettra des subventions allant jusqu'à 50% du coût du projet pendant les deux premières années. https://www.frenchweb.fr/blast-un-programme-soutenu-par-starburst-qui-veut-faire-exploser-les-startups-francaises-du-spatial/410612

  • Trustworthy AI: A Conversation with NIST's Chuck Romine

    21 janvier 2020

    Trustworthy AI: A Conversation with NIST's Chuck Romine

    By: Charles Romine Artificial Intelligence (AI) promises to grow the economy and improve our lives, but with these benefits, it also brings new risks that society is grappling with. How can we be sure this new technology is not just innovative and helpful, but also trustworthy, unbiased, and resilient in the face of attack? We sat down with NIST Information Technology Lab Director Chuck Romine to learn how measurement science can help provide answers. How would you define artificial intelligence? How is it different from regular computing? One of the challenges with defining artificial intelligence is that if you put 10 people in a room, you get 11 different definitions. It's a moving target. We haven't converged yet on exactly what the definition is, but I think NIST can play an important role here. What we can't do, and what we never do, is go off in a room and think deep thoughts and say we have the definition. We engage the community. That said, we're using a narrow working definition specifically for the satisfaction of the Executive Order on Maintaining American Leadership in Artificial Intelligence, which makes us responsible for providing guidance to the federal government on how it should engage in the standards arena for AI. We acknowledge that there are multiple definitions out there, but from our perspective, an AI system is one that exhibits reasoning and performs some sort of automated decision-making without the interference of a human. There's a lot of talk at NIST about “trustworthy” AI. What is trustworthy AI? Why do we need AI systems to be trustworthy? AI systems will need to exhibit characteristics like resilience, security and privacy if they're going to be useful and people can adopt them without fear. That's what we mean by trustworthy. Our aim is to help ensure these desirable characteristics. We want systems that are capable of either combating cybersecurity attacks, or, perhaps more importantly, at least recognizing when they are being attacked. We need to protect people's privacy. If systems are going to operate in life-or-death type of environments, whether it's in medicine or transportation, people need to be able to trust AI will make the right decisions and not jeopardize their health or well-being. Resilience is important. An artificial intelligence system needs to be able to fail gracefully. For example, let's say you train an artificial intelligence system to operate in a certain environment. Well, what if the system is taken out of its comfort zone, so to speak? One very real possibility is catastrophic failure. That's clearly not desirable, especially if you have the AI deployed in systems that operate critical infrastructure or our transportation systems. So, if the AI is outside of the boundaries of its nominal operating environment, can it fail in such a way that it doesn't cause a disaster, and can it recover from that in a way that allows it to continue to operate? These are the characteristics that we're looking for in a trustworthy artificial intelligence system. NIST is supposed to be helping industry before they even know they needed us to. What are we thinking about in this area that is beyond the present state of development of AI? Industry has a remarkable ability to innovate and to provide new capabilities that people don't even realize that they need or want. And they're doing that now in the AI consumer space. What they don't often do is to combine that push to market with deep thought about how to measure characteristics that are going to be important in the future. And we're talking about, again, privacy, security and resilience ... trustworthiness. Those things are critically important, but many companies that are developing and marketing new AI capabilities and products may not have taken those characteristics into consideration. Ultimately, I think there's a risk of a consumer backlash where people may start saying these things are too easy to compromise and they're betraying too much of my personal information, so get them out of my house. What we can do to help, and the reason that we've prioritized trustworthy AI, is we can provide that foundational work that people in the consumer space need to manage those risks overall. And I think that the drumbeat for that will get increasingly louder as AI systems begin to be marketed for more than entertainment. Especially at the point when they start to operate critical infrastructure, we're going to need a little more assurance. That's where NIST can come together with industry to think about those things, and we've already had some conversations with industry about what trustworthy AI means and how we can get there. I'm often asked, how is it even possible to influence a trillion-dollar, multitrillion-dollar industry on a budget of $150 million? And the answer is, if we were sitting in our offices doing our own work independent of industry, we would never be able to. But that's not what we do. We can work in partnership with industry, and we do that routinely. And they trust us, they're thrilled when we show up, and they're eager to work with us. AI is a scary idea for some people. They've seen “I, Robot,” or “The Matrix,” or “The Terminator.” What would you say to help them allay these fears? I think some of this has been overhyped. At the same time, I think it's important to acknowledge that risks are there, and that they can be pretty high if they're not managed ahead of time. For the foreseeable future, however, these systems are going to be too fragile and too dependent on us to worry about them taking over. I think the biggest revolution is not AI taking over, but AI augmenting human intelligence. We're seeing examples of that now, for instance, in the area of face recognition. The algorithms for face recognition have improved at an astonishing rate over the last seven years. We're now at the point where, under controlled circumstances, the best artificial intelligence algorithms perform on par with the best human face recognizers. A fascinating thing we learned recently, and published in a report, is that if you take two trained human face recognizers and put them together, the dual system doesn't perform appreciably better than either one of them alone. If you take two top-performing algorithms, the combination of the two doesn't really perform much better than either one of them alone. But if you put the best algorithm together with a trained recognizer, that system performs substantially better than either one of them alone. So, I think, human augmentation by AI is going to be the revolution. What's next? I think one of the things that is going to be necessary for us is pulling out the desirable characteristics like usability, interoperability, resilience, security, privacy and all the things that will require a certain amount of care to build into the systems, and get innovators to start incorporating them. Guidance and standards can help to do that. Last year, we published our plan for how the federal government should engage in the AI standards development process. I think there's general agreement that guidance will be needed for interoperability, security, reliability, robustness, these characteristics that we want AI systems to exhibit if they're going to be trusted. https://www.nist.gov/blogs/taking-measure/trustworthy-ai-conversation-nists-chuck-romine

  • Additive Technologies For Future UK Air Power Advance

    10 juillet 2020

    Additive Technologies For Future UK Air Power Advance

    Tony Osborne July 10, 2020 Bidders pushing for a role to build a technology demonstrator for the UK's Lightweight Affordable Novel Combat Aircraft (LANCA) are waiting to see if their design proposals will be approved for the next phase of the initiative. Three industry teams were selected last summer (AW&ST July 29-Aug. 18, 2019, p. 18) to take forward development of the LANCA air system, an unmanned air vehicle that could act as additive capability accompanying future combat aircraft into operations. LANCA would perform a range of tasks, including serving as a loyal wingman, gathering intelligence or acting as a weapons carrier. The Royal Air Force envisions a platform costing 1/10th of current combat aircraft and developed in one-fifth of the time. The idea is like that of the Kratos XQ-58 Valkyrie for the U.S. Air Force's Skyborg program, Australia's Boeing-led Airpower Teaming System and remote carriers for the French, German and Spanish Future Combat Air System. Since then, the industry teams—which include Boeing Phantom Works International, working with Marshall Aerospace and Defence Group and Cranfield University; consortium Team Black Dawn, Callen-Lenz and Bombardier Belfast; and Northrop Grumman UK with Team Avenger, led by Blue Bear Systems Research and yet to be disclosed partners—submitted their proposals for the £4.8 million ($6 million) Phase 1. The LANCA program is being led by the UK Defense Science and Technology Laboratories in conjunction with the Royal Air Force's Rapid Capability Office. The UK Defense Ministry is evaluating proposals in readiness for the second phase of the project, called Mosquito, which is worth £30-50 million. Mosquito would see one, possibly two of the candidates being matured into a full-size flightworthy demonstrator potentially undertaking a flight-test program. Aviation Week has been told there was stiff competition for the LANCA program: Some 45 bidders entered Phase 1 and were subsequently scaled down to nine. Around eight bids were then tendered, of which three were chosen. Few details have emerged about the proposals, although Boeing Australia confirmed through social media that it had secured a “first-of-type permit” from the Australian government to share design material for its Airpower Teaming System with the UK, suggesting the Boeing proposal may borrow heavily from that platform. Progress is also occurring on a drone swarm system announced by former Defense Secretary Gavin Williamson in February 2019: Demonstrations in March using five unmanned air vehicles proved a collaborative capability between the platforms, people close to the program told Aviation Week. https://aviationweek.com/ad-week/additive-technologies-future-uk-air-power-advance

Toutes les nouvelles