CARP Mobile Sensing Framework in Flutter

pub package style: effective dart github stars MIT License arXiv

This library contains the core Flutter package for the CARP Mobile Sensing (CAMS) framework. Supports cross-platform (iOS and Android) mobile sensing.

For an overview of all CAMS packages, see CARP Mobile Sensing in Flutter. For documentation on how to use CAMS, see the CAMS wiki.


To use this plugin, add carp_mobile_sensing as dependencies in your pubspec.yaml file.

    sdk: flutter
  carp_core: ^latest
  carp_mobile_sensing: ^latest

Android Integration

The pedometer (step count) probe needs permission to ACTIVITY_RECOGNITION. Add the following to your app's manifest.xml file located in android/app/src/main:

    <uses-permission android:name="android.permission.ACTIVITY_RECOGNITION"/>

iOS Integration

The pedometer (step count) probe uses NSMotion on iOS and the NSMotionUsageDescription needs to be specified in the app's Info.plist file located in ios/Runner:

  <string>Collecting step count.</string>

NOTE: Other CAMS sampling packages require additional permissions in the manifest.xml or Info.plist files. See the documentation for each package.


The Dart API doc describes the different libraries and classes.

The wiki contains detailed documentation on the CARP Mobile Sensing Framework, including the domain model, how to use it by create a Study configuration, how to extend it, and an overview of the different Measure types available.

A more scientific documentation of CAMS is available at

  • Bardram, Jakob E. "The CARP Mobile Sensing Framework--A Cross-platform, Reactive, Programming Framework and Runtime Environment for Digital Phenotyping." arXiv preprint arXiv:2006.11904 (2020). [pdf]
  title={The CARP Mobile Sensing Framework--A Cross-platform, Reactive, Programming Framework and Runtime Environment for Digital Phenotyping},
  author={Bardram, Jakob E},
  journal={arXiv preprint arXiv:2006.11904},

Please use this as a reference in any scientific papers using CAMS.

Examples of configuring and using CAMS

There is a very simple example app app which shows how a study can be created with different tasks and measures. This app just prints the sensing data to a console screen on the phone. There is also a range of different examples on how to create a study to take inspiration from.

However, the CARP Mobile Sensing App provides a MUCH better example of how to use the framework in a Flutter BLoC architecture, including good documentation of how to do this.

Below is a small primer in the use of CAMS.

Following carp_core, a CAMS study can be configured, deployed, executed, and used in different steps:

  1. Define a StudyProtcol.
  2. Deploy this protocol to a DeploymentService.
  3. Get a study deployment for the phone and start executing this study deployment using a SmartPhoneClientManager.
  4. Use the generated data locally in the app or specify how and where to store or upload it using a DataEndPoint.

Note that as a mobile sensing framework running on a phone, CAMS could be limited to support 3-4. However, to support the 'full cycle', CAMS also supports 1-2. This allows for local creation, deployment, and execution of study protocols (which in many applications have shown to be useful).

Defining a StudyProtcol

In CAMS, a sensing protocol is configured in a StudyProtocol. Below is a simple example of how to set up a protocol that sense step counts (pedometer), ambient light (light), screen activity (screen), and power consumption (battery).

// import packages
import 'package:carp_core/carp_core.dart';
import 'package:carp_mobile_sensing/carp_mobile_sensing.dart';

void example() async {
  // create a study protocol storing data in files
  SmartphoneStudyProtocol protocol = SmartphoneStudyProtocol(
    ownerId: 'AB',
    name: 'Track patient movement',
    dataEndPoint: FileDataEndPoint(
      bufferSize: 500 * 1000,
      zip: true,
      encrypt: false,

  // Define which devices are used for data collection.
  // In this case, its only this smartphone.
  Smartphone phone = Smartphone();

  // Add a background task that immediately starts collecting step counts,
  // ambient light, screen activity, and battery level.
          Measure(type: SensorSamplingPackage.PEDOMETER),
          Measure(type: SensorSamplingPackage.LIGHT),
          Measure(type: DeviceSamplingPackage.SCREEN),
          Measure(type: DeviceSamplingPackage.BATTERY),

The above example defines a simple SmartphoneStudyProtocol which will store data in a file locally on the phone using a FileDataEndPoint. Sampling is configured by adding a TriggeredTask to the protocol using an ImmediateTrigger which triggers a BackgroundTask containing four different Measure.

Sampling can be configured in a very sophisticated ways, by specifying different types of devices, triggers, tasks, measures and sampling configurations. See the CAMS wiki for an overview.

You can write your own DataEndPoint definitions and coresponding DataManagers for uploading data to your own data endpoint. See the wiki on how to add a new data manager.

Using a DeploymentService

A device deployment specifies how a study protocol is executed on a specific device - in this case a smartphone. A StudyProtocol can be deployed to a DeploymentService which handles the deployment of protocols for different devices. CAMS comes with a simple deployment service (the SmartphoneDeploymentService) which runs locally on the phone. This can be used to deploy a protocol and get back a MasterDeviceDeployment, which can be executed on the phone.


// Use the on-phone deployment service.
DeploymentService deploymentService = SmartphoneDeploymentService();

// Create a study deployment using the protocol
StudyDeploymentStatus status =
    await deploymentService.createStudyDeployment(protocol);

Running a SmartphoneDeploymentController

A study deployment for a phone (master device) is handled by a SmartPhoneClientManager. This client manager controls the execution of a study deployment using a SmartphoneDeploymentController.


String studyDeploymentId = ... // any id obtained e.g. from an invitation
String deviceRolename = ... // the rolename of this phone in the protocol;

// Create and configure a client manager for this phone
SmartPhoneClientManager client = SmartPhoneClientManager();
await client.configure(deploymentService: deploymentService);

// Create a study object based on the deployment id and the rolename
Study study = Study(studyDeploymentId, deviceRoleName);

// Add the study to the client manager and get a study runtime to control this deployment.
await client.addStudy(study);
SmartphoneDeploymentController? controller = client.getStudyRuntime(study);

// Deploy the study on this phone.
await controller?.tryDeployment();

// Configure the controller and start sampling.
await controller?.configure();

Using the generated data

The generated data can be accessed and used in the app. Access to data is done by listening on the data streams from the study deployment controller or some of its underlying executors or probes. Below are a few examples on how to listen on data streams.


// listening to the stream of all data events from the controller => print(dataPoint));

// listen only on CARP events
    .where((dataPoint) =>!.format.namespace == NameSpace.CARP)
    .listen((event) => print(event));

// listen on LIGHT events only
    .where((dataPoint) =>!.format.toString() == SensorSamplingPackage.LIGHT)
    .listen((event) => print(event));

// map events to JSON and then print
    .map((dataPoint) => dataPoint.toJson())
    .listen((event) => print(event));

// subscribe to the stream of data
StreamSubscription<DataPoint> subscription =
    controller!.data.listen((DataPoint dataPoint) {
  // do something w. the datum, e.g. print the json
  print(JsonEncoder.withIndent(' ').convert(dataPoint));

Controlling the sampling of data

The execution of sensing can be controlled on runtime in a number of ways. For example:


// Sampling can be paused and resumed

// Pause specific probe(s)
  .forEach((probe) => probe.pause());

// Adapt a measures.
// Note that this will only work if the protocol is created locally on the
// phone (as in the example above)
// If downloaded and deserialized from json, then we need to locate the
// measures in the deployment
  ..overrideSamplingConfiguration = PeriodicSamplingConfiguration(
    interval: const Duration(minutes: 5),
    duration: const Duration(seconds: 10),

// Restart the light probe(s)
    .forEach((probe) => probe.restart());

// Alternatively mark the deplyment as changed - calling hasChanged()
// this will force a restart of the entire sampling

// Once the sampling has to stop, e.g. in a Flutter dispose() methods, call stop.
// Note that once a sampling has stopped, it cannot be restarted.
await subscription.cancel();

Features and bugs

Please read about existing issues and file new feature requests and bug reports at the issue tracker.


This software is copyright (c) Copenhagen Center for Health Technology (CACHET) at the Technical University of Denmark (DTU). This software is available 'as-is' under a MIT license.


This library contains the software architecture for the CARP Mobile Sensing (CAMS) framework implemented in Flutter. Supports cross-platform (iOS and Android) sensing.
A library containing a sampling package for collecting information from the device hardware:
The CAMS implementation of the core CARP domain classes like StudyProtocol, TaskDescriptor, and Measure. Also hold JSON serialization and deseralization logic to handle seraialization of the domain objects.
A library for data managers. Contains implementation of the
Contains classes for running the sensing framework incl. the StudyDeploymentExecutor, TaskExecutor and different types of Probes.
A library containing a sampling package for collecting data from the basic device sensors: