CarefulAI
CEOs Feed
  • AI Assurance
  • AI Model Design
  • Motives and Models
  • Model Feedback
  • About Us

AI Model Design

Our Values: Users should set targets for accuracy, sensitivity and specificity, these should take priority over academic and industrial priorities.

Our mission is to enable customers to design ML and AI that:


​1. Enables a tenfold improvement in a defined user experience
2. Protects the privacy and rights of humans.
3. Mitigates technical and clinical risks

4. Can be freely released when it is in the public interest

​We maintain strong co-development partnerships with:
Picture
Recent models and tool design includes:

  • Triaging breathlessness in call recordings (ARDAR)
  • Mitigate suicide risk in text supplied via online forms (SISDA)
  • Reduction of social isolation amongst Carers (CAREDAR)
  • Improving the delivering of remote training in the area of patient engagement (EMPDAR)
  • Annotating mobile images for use in skin cancer classifier development (SKINDAR)
  • Live ethical monitoring of population health e.g. Wellbeing Maps

For more detail see 'Motives and Models'
​

The 'DAR Design Framework': A Unique Ethical AI Toolset

​​We encourage the use of these privacy protecting development tools:
​
  • ​PRIDAR: a method of prioritising the ethical and safe development of AI solutions
  • VOICEDAR AND GAPDAR: enable users to draw an inference from human-to-human and human-to-computer interaction
  • WHATDAR: enables users to gather insight from unstructured audio and text conversations
  • NUDGESHARE: an ethical method of gathering self reported data from the public

These tools enable AI powered models to be deployed globally
Picture

Our Professional Practice


We are committed to User Centred Design:

  • We help customers understand their role within the AI funding, users and supplier ecosystem
  • Support users and subject matter experts committed to the development of ethical AI: via user group moderation and insight development. To join a group, or to form a new group contact us
  • We are committed to continuously gathering and acting upon model feedback

We are committed to Privacy by Design

  • By the implementation of good practice as defined by the ICO
  • Designing systems that do not gather personally identifiable data
  • Designing systems that put users in control of the data they share
  • Enabling CarefulAI System and Model users to share Data with others using informed explicit consent

We are committed to managing the risks and rewards associated with AI design. To this end we use

  • PRIDAR: a method for prioritising risk associated individual AI driven care solutions
  • VALDAR: an approach to monitoring the value and safety of AI in a pathway​

We help customers to apply the required standards and procedures:

  • NICE, MHRA, ICO, MDR, EU and FDA Regulations
  • DCB0129/160
  • ISO 9001, 13485, 27001, 27701
  • IEC 62304
  • Purchasing standards e.g. DTAC

​We encourage customers to apply design best practice:​

  • Managing the data Pipeline and Production using LDP & Federated ML to GDS Standards
  • Audit data inputs and outputs and look for subgroup bias
  • Apply simple design & sanity & fairness checks on external data sources
  • Use adaptive stress techniques to look for model weaknesses e.g. Adastress
  • Employ interpretable models in high risk environments
  • Peer review training scripts
  • Avoid external dependencies
  • Automate feature, parameter and model selection and configuration
  • Ensure all features have an owner and a well documented rationale
  • Version models and data sources
  • Support Model Card Reporting and AI Assurance e.g. ​MCT , CAF
  • Using 'Red Teams' and GANs to identify system weaknesses e.g. Infodemics

We have a best in breed deployment stack:

  • Commit to in-country and partner standards e.g. EU , Google , Microsoft, IBM
  • Adapt Tooling to the User Needs
  • Building Verifiable Data Logs e.g. Trillian
  • Effective Code Deployment i.e. Travis
  • Automated Regression Tests i.e. JUnit, Python
  • Virtualise model deployment utilising LDP and Federated Learning
  • Automating Model Deployment i.e. Kubeflow and Docker
  • Continuously measure model performance

​We enable customers and partners to be transparent in how models are licensed

  1. With the agreement of NHS customers, Generalised ML / AI models derived from NHS work are made available gratis to other NHS customers (see Terms of Use)
  2. Models that cannot be released freely are available as supported products for specific user cases under licence, for which there is a monthly fee. The terms under which models are released are shown in our Licence Builder. The licence builder can be used to propose the terms under which you wish to apply a model.

London Office

What Our Clients Are Saying

''I love your mantra.  AI is Artificial Ignorance until it is validated by people for people.' NHSE Director

'You deliver technically complex subjects beautifully' NHSE Clinician

'Federated AI powered by a mobile phone, creating insights that are useful to the user, which they then control .  Brilliant.' NHSE Clinician

Picture