ICYMI: AI is your coworker and cybersecurity needs a nudge
AI roadmaps: 60% of UK businesses struggle for analytics ROI
Analytics Automation company Alteryx has uncovered the core factors blocking organisations from delivering in-demand business intelligence with data analytics. New Alteryx-commissioned IDC research highlights that - despite businesses needing faster, more accurate insights to navigate competitive markets, businesses are still attempting to meet their future AI analytic goals with legacy strategy.
The research showcases the negative ROI impact of lagging upskilling, democratisation, and accessibility initiatives. When these foundational data pillars are not prioritised, the result is a broken link between in-department expert and analytic-driven insight - one where analytics is limited to those who can code and not the person best placed to answer the question.
Read the full story here.
AI is a coworker, not a job threat, say business leaders
A substantial majority of managers and executives regard artificial intelligence (AI) as a coworker, not a job threat, according to new research. A new report from MIT Sloan Management Review (MIT SMR) and Boston Consulting Group (BCG) found that 60 per cent of employees have welcomed AI as a coworker, and organisations with employees working well with AI are 5.9 times more likely to see significant financial benefits than organisations that are not using AI.
Read the full story here.
New AI research could drive progress in clean energy
Meta has revealed what it describes as an AI breakthrough that accelerates protein folding — the process of predicting a molecule’s shape. Meta researchers have used it to create a database of the molecular structures of proteins, which are the building blocks of life on earth.
Read the full story here.
Cybersecurity may fail without nudge in the right direction
Most employees would try to circumvent security controls that block access to unsanctioned applications at work, cybersecurity research has revealed, but a more positive experience could improve matters.
According to new research from Nudge Security, undesirable security behaviours may have less to do with a lack of awareness and more to do with basic human emotions. The company’s new report, Debunking The 'Stupid User' Myth in Security, explores how workers' attitudes and emotions influence security behaviours.
Read the full story here.
AI strategies must become beating heart of security measures
Artificial Intelligence-enabled strategies must become the beating heart of security measures, if organisations wish to protect their identities, both human and machine, from evolving cyber threats, after the National Cyber Security Centre revealed that it had issued 34 million cyber alerts in the past year alone. The agency’s NCSC Annual Review 2022 report, published this week, details its progress in helping to turn the UK into one of the world’s safest places to live and work online.
Read the full story here.
Reality not required as synthetic data powers machine models
Researchers currently train machine-learning models using large datasets of video clips showing humans performing actions. However, not only is this expensive and labour-intensive, but the clips often contain sensitive information. Using these videos might also violate copyright or data protection laws as many datasets are owned by companies and are not available for free use.
To address this, a team of researchers at MIT, the MIT-IBM Watson AI Lab, and Boston University built a synthetic dataset of 150,000 video clips that captured a wide range of human actions, which they used to train machine-learning models. These are made by a computer that uses 3D models of scenes, objects, and humans to quickly produce various clips of specific actions — without the potential copyright issues or ethical concerns that come with real data.
Read the full story here.