Young’s age and background – things that might seem like disadvantages when it comes to more established industries – have become his secret weapons. When he walks into a room of executives twice or three times his age, he says, there’s initial skepticism. “Who the hell is this young guy and how does he know what he’s talking about?”
Artificial intelligence helps patients better comprehend medical findings
Medical reports written in technical terminology can pose challenges for patients. A team at the Technical University of Munich (TUM) has investigated how artificial intelligence can make CT findings easier to understand.
How can you tell if music is AI-generated?
As AI-generated music floods streaming platforms, questions bubble over whether listeners are owed more transparency.
N.Y. Assemblyman Clyde Vanel on Trump’s possible AI executive order
A Trump order would block states from enforcing their own regulations around artificial intelligence.
Google denies ‘misleading’ reports of Gmail using your emails to train AI
Google says “we do not use your Gmail content for training our Gemini AI model.”
Development and validation of an artificial intelligence-based model for cardiovascular disease prediction using longitudinal data – BMC Medical Informatics and Decision Making
Background This research uses machine learning (ML) models to determine significant predictive factors for CVD events in Iran, a country with high mortality rates. The study evaluates the effectiveness of deep learning and mixed-effects logistic models in predicting 10-year CVD incidence using longitudinal TLGS data. Methods A total of 4,872 adults aged ≥ 30 years without a history of CVD at baseline (2006–2008) were followed until 2020. Following exclusions due to prevalent CVD or insufficient follow-up data, the final analytic sample comprised 1,942 men and 2,930 women. Baseline demographic, behavioral, and biochemical characteristics were utilized as input features. A clinical history examination was employed to identify and confirm incident CVD events, such as stroke and coronary heart disease (CHD). Using longitudinal data collected during a 10-year study period, deep learning models based on Long Short-Term Memory (LSTM) and Gated Recurrent Unit (GRU) architectures were utilized to detect variations in risk factor dynamics. The area under the receiver operating characteristic curve (AUC) was employed to assess the model’s performance. Results During the follow-up period, 545 participants (11.2%) experienced CVD events. In both sexes, the GRU model outperformed the LSTM model. In women, the GRU achieved an AUC of 0.739, whereas in men it achieved 0.738, in both cases outperforming LSTM. Compared with the traditional mixed-effects model, discrimination was comparable in women (GRU 0.739 vs. 0.74) and higher in men (0.738 vs. 0.70). Even though the models were restricted to 21 commonly employed clinical variables, their performance was similar to that of larger-scale studies that included hundreds of variables. Conclusions Deep learning models, such as GRU, can effectively leverage longitudinal medical data to predict future cardiovascular disease events. The models achieved performance comparable to studies that used far more variables, despite relying on a limited feature set.
From Pen to Portal: Using Artificial Intelligence to Evaluate the Impact of Electronic Patient Information and Communication (EPIC) on the Quality of Urology Clinic Letters at a Major London Hospital
Background
High-quality documentation is an essential part of providing patients with good clinical care. Clinical letters promote continuity, reduce errors, and communicate effectively betwe…
On AI, States Generally Seek Innovation With Protection
A state-by-state AI policy scan from the Council of State Governments offers a clear and comparative view of the AI governance landscape across the U.S., even as the federal government eyes restrictions.
Bret Taylor’s Sierra reaches $100M ARR in under two years
The startup’s rapid growth suggests that enerprises are embracing AI agents.
Microsoft's head of AI doesn't understand why people don't like AI, and I don't understand why he doesn't understand because it's pretty obvious
Is it really “mindblowing” that people are skeptical of software that consistently doesn’t do the things we’re told it can do?
