Loading

kelleni2 0

Activity

May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
Jan
Feb
Mar
Apr
May
Mon
Wed
Fri

Ratings Progression

Loading...

Challenge Categories

Loading...
Gold 0
Silver 0
Bronze 0
kelleni2 has not joined any teams yet...

Novartis DSAI Challenge

DSAI Challenge Evaluation - next steps

4 months ago

Please note

  •   The date and location for the Challenge Event has not yet determined
    
  •   Travel cannot be covered by the data challenge team
    
  •   Please talk to your manager if travel can be supported, in particular if you belong to the finalists and potential awardees
    
  •   Details will be shared as soon as possible

DSAI Challenge Evaluation - next steps

4 months ago

Dear all,

We wanted to share the upcoming milestones regarding challenge evaluation:

  1. Next 2 weeks - Core challenge team to shortlist top solutions for subjective categories and validate top performing solutions from the leaderboard:
  •      Technical evaluation of top solutions – robustness, generalizability
    
  •   	Evaluation of innovative approaches and business value for the respective categories
    
  1.  Broader evaluation committee to confirm final selections – potentially reaching out during this process. 
    
  2.  Finalists to be notified mid-March and prepped to present their solutions at the Challenge event in front of the committee, business, and community – and compete for best presentation!
    
  3.  All participants are invited to the Challenge Event in Q2 to exchange learnings on everything from PoS, to platform feedback, to “what’s next” for future challenges at Novartis. 
    

The team will be reaching out with any questions or requests regarding your solution. We look forward to future communications around the event and seeing you there!

Best,
Nick and core challenge team

Urgent: Request asap if you need access to your vm this week - vm's start being deleted Tuesday morning

5 months ago

Dear all,

You will have access to your workspaces throughout this week.

If you need continued access to your VM - which is launched from the workspace- you can request this.

We will delete all other VM’s starting tomorrow but we can make exceptions where needed.

Please request below by replying to this thread.

In either case, please make sure you have followed directions to transfer your VM files to the workspace so you do not lose your VM data/files once the VM’s are deleted.

Best,
Nick

DSAI Challenge UPDATE: decommissioning workspaces, final presentations, leaderboard

5 months ago

Hi Shravan,

I had to read your question more carefully - but yes you have access to the code etc via gitlab and can use your local machine to prepare, and push back to gitlab. If you have trouble here - don’t stress.

Best,
Nick

REMINDER: data in the VM's will be deleted starting Monday 13th

5 months ago

Hi all,

Reminder that we plan to start the process permanently deleting the VM’s Monday 13th.

The workspaces they are launched from will remain operational through the end of next week.

Please reach out to Ned with any questions regarding the workspaces, and see this post regarding where to place data/files that you would like to keep.

DSAI Challenge UPDATE: decommissioning workspaces, final presentations, leaderboard

5 months ago

Hi - some quick answers:

Yes on 1. and 2.

Yes you can utilize time until 13th to clean up code etc

The hard deadline is when the VM’s will start being deleted - but we will have a call monday morning before this in case difficulties reported.

The workspaces will be available next week, but the VM’s launched from them will not be. I’ll post again more detail on the presentations. (but it will be certainly be possible to modify them next week)

DSAI Challenge UPDATE: decommissioning workspaces, final presentations, leaderboard

5 months ago

Dear all,

Based on feedback from teams as well as the evaluation team and process, we are happy to extend the presentation deadline and keep the workspaces open through the end of next week.

  • Presentations submitted by end of week Jan 17.
  • Code and models due Friday 10th – hard deadline “before January 13th” as the evaluation team will start

(note: VM’s will still go down)

Best,
Nick

DSAI Challenge Update: Presentations can be submitted through week Jan 17

5 months ago

Dear all,

Based on feedback from teams as well as the evaluation team and process, we are happy to extend the presentation deadline and keep the workspaces open through the end of next week.

  • Presentations submitted by end of week Jan 17.
  • Code and models due Friday 10th – hard deadline “before January 13th” as the evaluation team will start

(note: VM’s will still go down)

Best,
Nick

DSAI Challenge UPDATE: decommissioning workspaces, final presentations, leaderboard

5 months ago

Dear all,

Happy new year and welcome back!

As the Solutioning phase of the DSAI Challenge comes to a close – I wanted to make sure we were aware of the process – and if there existed any questions, concerns or general panic:

Virtual Machines will be brought down starting Jan 13. All valuable files and data must be copied to the appropriate shared location or it will be lost forever.

  • See here
  • Please reach out now on the forums if there are concerns!

Leaderboard:

  • Public view locked on 10th – EoB (6pm) EST
  • Submissions will be re-run including the hold-out data. Your best submission will count – based on monitoring and discussions with AI Crowd we have not seen “gaming” with frequent submissions
  • Therefore, Winning solutions will be under some scrutiny – i.e. if it looks like a statistical fluctuation from many submissions gave an edge, this will be taken into account and adjusted.
  • And will consider other factors as well to differentiate between and acknowledge multiple top solutions.

Presentations should be ideally be placed into GIT by EoB Friday.

  • Please reach out if you are having trouble or if you have questions with presentations or some extenuating circumstance.
  • Please email me directly – some short grace period could be possible if we know in advance.
  • Also we can post about getting files into git – don’t stress
  • In general - are teams now ready with their presentations??

Hope everyone is feeling good about their solutions and presentation – looking through the code and discussions, it seems there are quite some interesting approaches!

Best,
Nick and team

UPDATE / EXTENSION: DSAI Challenge: Leaderboard & Presentation deadlines

5 months ago

Hi Bin,

I am sending an update now - as suggested in the last update - we are no longer limited by last submission, so do not worry there. We have not observed gaming and have ability to rerun on multiple now - see the coming announcement.

Jan 10th EoB EST is what we had planned - but open to feedback. (will give precise time)

IMPORTANT INFORMATION - Presentation work in workspaces

5 months ago

Hi Shravan -
I’m sending an update now. In theory we had specified the submitted presentation timeline was also 10th. Please email me if you have concerns about the deadline, open to feedback.
Best,
Nick
nicholas.kelley@novartis.com

Reminder: All code must be checked into GIT to validate Leaderboard submissions

5 months ago

Hi all,

First off - for the most part repositories and reproducibility of solutions are looking good!

Just a reminder that to be eligible - all code surrounding the various aspects of what was needed to enable your model’s Leaderboard predictions must be submitted to GIT. Top performing models will need to be validated/interrogated by team (information leak, etc) as well as ensuring we have reproducible solutions.

“Special circumstances”, uncertainty or questions if having trouble - just reach out to Shivam and the AI Crowd team - they will be looking to help!

Thanks!
Nick

UPDATE / EXTENSION: DSAI Challenge: Leaderboard & Presentation deadlines

5 months ago

Dear all,

IMPORTANT UPDATE: We have made some necessary changes to the test data set which will allow for better balance and stability as we transition to the final leaderboard, and may cause the current leaderboard to shuffle. (models will be re-run)

As a consequence of this – in addition to some general requests - we assume it is only fair to allow teams an additional non-holiday week to work on finalizing their submission.

  1. Leaderboard submissions will be frozen on Jan 10th
  2. The organizing team will decide if an additional grace period will be given for presentation submissions.
  3. Machines are accessible from now until Jan 10th

As always, thanks for your continued engagement and support of the organizing team’s efforts around your feedback throughout the initiative.

Thanks and happy holidays

Challenge timelines updates/clarifications

5 months ago

Dear all,

We just wanted to clarify a few things on the upcoming Challenge process:

  1. Leaderboard submissions will be frozen on Dec 20th – no further submissions will be possible

    • Your final submission will then be used on the holdout data – please reach out to Satya if this causes problem and we will adjust
  2. Final presentations can be submitted the week of Jan 6th. Due to perceived need based on questions, guidance coming in a separate mail tomorrow morning.

  3. Workspaces will be unavailable starting Dec 23rd until Jan 6th but will be accessible JAN 6th UNTIL JAN 10TH. (thanks Aridhia).

    • This will allow you to get data and relevant files off of them, as well as to upload your final presentation to git if you have not already done so.
  4. Once the workstations are decommissioned Jan 10th all data and code on them will disappear EXCEPT:

    • SOLUTIONS AND PRESENTATION MUST BE IN AICROWD GIT REPOSITORY
      i. code/notebooks should happen somewhat automatically if you participated with the leaderboard – PLEASE REACH OUT IF YOU ARE UNSURE.

    • ALL NEW WRANGLED DATA OR FILES SHOULD BE TRANSFERRED TO THE COMMON SPACE – /home/workspace/files/teams/your_team_name – otherwise it will be lost. Copy it over so we can all benefit and you’ll be eligible for wrangling recognition.

Please let Satya and/or myself know if there are major concerns.

Presentation guidance/suggestions (due Jan)

5 months ago

Some guidance on presentations due in January – this will live in a live version on the forums should details/suggestions need to be added.

Here are some general tips for what to communicate:

  1. Explain what you did for a general audience - clarity, transparency, scientific integrity of information

  2. “Confidence in the model”: Convince us that your model performs and is generalizable and robust – especially if you used additional data so that we are sure your performance was not due to information leak.

  3. “Model transparency”: To the extent possible, provide insight or transparency on how your model is making predictions – top features, how the features are related or their context, etc

  4. “Innovation”: Your top most innovative aspects you focused on either Data Wrangling or Novel methodology / approach

  5. “Insights”: What did we learn at a high level about the risks and success factors of trials, etc? (possibly covered in other sections)

  6. “Enable business decisions”: Make sure to connect any actual insights or tools/methods that a clinical project team might be able to act on or learn from – or simply see the problem through a different lens.

For example: consider how your model could be used in the context of various decisions – in the context of a portfolio for example, where we might want to understand the degree of correlation multiple assets might be in their outcomes.

Please review the full list of proposed insights to provide:
https://www.aicrowd.com/organizers/novartis/challenges/novartis-dsai-challenge#detailed-evaluation-criteria

Final submission selection

5 months ago

Hi all,

We will use the last submission.

This will allow users to decide which submission they feel is their best.

If this becomes problematic based on feedback, we could also give the option to use the submission with the best score. Email incoming today as well about the challenge end - but please comment below.

Best,
Nick

New data available - structured phase 3 information

6 months ago

Thanks so much Ned!

And thanks to the Hyderabad team who generously provided this wrangled data set in order to help encourage additional explorations into this portion of the insights challenges!

Reminder to particpants:

YOU MAY NOT USE PHASE 3 DATA IN THE LEADERBOARD PREDICTIONS.

The simple existence of a phase 3 being launched carries information - as well as much of it not being available at the time of the phase 3 investment decision being made.

HOWEVER - PLEAES DO EXPLORE THE INFORMATION IT CARRIES FOR POST-PHASE 2 PREDICTIONS FOR THE INSIGHTS SECTION!

good luck, nick and team

Submission of only final predictions file

6 months ago

It would help to go back to the underlying motivation:

  • We wanted to reduce the chance of visibly fooling ourselves with top solutions including leaked information, rendering them irrelevant for real world decision making.

  • We wanted all top solutions able to be re-run by the evaluation & project team, to be interrogated for generalizability etc. By design, the kubernetes cluster and git combo enables this.

That said, we also want the best solutions possible for the larger initiative at the end of the event - which is why we were trying to ease some of the frustrations which were blocking some teams.

I discussed with the team, and we would highly encourage to continue to predict on the original test data in the evaluation clusters rather than provide a table of solutions. Especially for the final solution.

However, do what you feel you need to do as a team in order to come up with your optimal solution. But keep in mind, the final leaderboard will change when we add in the hold out test data, and winners will need their model to be validated by the evaluation team, so please make it clear how one would load and interrogate your model.

Test data matrix available

6 months ago

Please always feel free to reach out to me if you would like to discuss.

Test data matrix available

6 months ago

Hi Bjoern,

It would help to go back to the underlying motivation:

  • We wanted to reduce the chance of visibly fooling ourselves with top solutions including leaked information, rendering them irrelevant for real world decision making.
  • We wanted all top solutions able to be re-run by the evaluation & project team, to be interrogated for generalizability etc. By design, the kubernetes cluster and git combo enables this.

That said, we also want the best solutions possible for the larger initiative at the end of the event - which is why we were trying to ease some of the frustrations which were blocking some teams.

I discussed with the team, and we would highly encourage to continue to predict on the original test data in the evaluation clusters rather than provide a table of solutions. Especially for the final solution.

However, do what you feel you need to do as a team in order to come up with your optimal solution. But keep in mind, the final leaderboard will change when we add in the hold out test data, and winners will need their model to be validated by the evaluation team, so please make it clear how one would load and interrogate your model.

1 week exrtension

6 months ago

Dear all,

The project team has suggested that we allow an additional week for those who need it for final submissions given some of the initial technical challenges.

Please submit your final insights and model before the Christmas holidays - Friday 20th.

Best,
Nick

Any information on how the final insights should be submitted?

6 months ago

There are different aspects. The leaderboard can be though of as the methods competition, and due to the way you submit, your model and code will already be visible in gitlab. Additional evidence can be provided there as well in your code or notebook, to demonstrate that no information was leaked and your model is generalizable – especially if you used additional data.

The insights part of the challenge would require communicating your learnings. This could be in a jupyter notebook, or powerpoint, or however you would like. That should be submitted also in December - and also uploaded to gitlab. Finalists will present in January to our evaluation committee. I’ll post who this will consist of soon, but it will be a mix of AI, problem owners, and data strategy.

Is the scoring function F1 or logloss?

6 months ago

hi bjoern - right now it is the best log loss submission.

please keep in mind that in the test data - we do have a hold out.

the final leaderboard will be the hold out test data - plus rthe current test data. this would be evaluated currently on your top submitted model.

Duplicates of intClinicalTrialID

6 months ago

hi - yes many columns including trial ID are not unique

a trial can have multiple indications officially

a clinical “program” which we trying to predict is the approval of a drug-indication pair - and that should always have the same label in the data.

Test data matrix available

6 months ago

The test data matrix is now available to ease some logistical reasons.

IMPORTANT CONSIDERATIONS SHOULD YOU CHOOSE TO ACCESS THIS FILE:

  • The leaderboard is meant to emulate real world decision making scenarios where you would only have access to information in the past – immediately following your phase 2. Please keep this in mind when if attempting to leverage the test data for any learning. Please justify any choices and show generalizability of your model if leveraging additional data from the post-2016 test data set.

As with wrangling raw data - Please Consider: Solutions and predictions by teams choosing to leverage additional data for the leaderboard methods competition beyond the core training data set will be under an additional layer of oversight in both code and model generalizability to ensure no information has been leaked.

For now, find this file in your region’s respective /shared_data/data directory in test_data_full

RAW DATA now available in shared folder

6 months ago

Dear all,

The team has now place the raw MIT/Informa tables for our data wranglers to explore in the /shared_data folders under “raw”. We can start a thread from our data wranglers if there are questions on the data.

As of now, and explained earlier for a variety of reasons, this data is not available in this raw format to the evaluation engine.

Since we will now be providing the test data, you will still be able to wrangle raw data to the test data and pass along additional columns to the evaluator.

NOTICE: It is the responsibility of the team to ensure that any additional data:

  • DOES NOT CONTAIN INFORMATION AFTER 2015
  • DOES NOT CONTAIN INFORMATION ABOUT THE PHASE 3 TRIAL FOR A LEADERBOARD PREDICTION – note that this is actually encouraged for some of the challenge insights questions!
  • Which obviously includes, but is not limited to the outcome of the trial itself

Please Consider: Solutions and predictions by teams choosing to add additional data will be under an additional layer of oversight in both code and model generalizability to ensure no information has been leaked.

If you have observed a performance increase from specific data and would like to pass this into the evaluation cluster – making this available to other participants and to receive better validation that it is not leaking information, this sharing would be recognized especially if demonstrated to enable the competition and is highly encouraged - please email me or post on the forums.

Original Datasets for Train and Test

6 months ago

The test data was originally not intended to be visible other than a sample file for column names and format.

However, we will plan to make the test data available due to various logistical reasons for those who feel they need it. I will create a separate post on that topic.

How to use conda-forge or CRAN for packages in evaluation?

6 months ago

hi @shivam - could you see if you could help Bjoern? And adding the learnings to the FAQ or forums would be great.

@bjoern.holzhauer - you need details on the docker etc? I can connect the two of you via email if it helps.

Drug ID conversion tables and code

6 months ago

Dear all,

As discussed on the Informa API walkthrough, here are the drugID conversions where possible into things like smiles, NVP, CAS, etc

You can find the tables and some sample code used in it’s creation in:

Please note that this currently contains mappings for both the training and test sets.

Intphaseendyear 1900

6 months ago

From Imran at Informa:
“Usually when you see a date of 1900-01-01 it means that we do not know the date.”

He also said he can investigate but we will have the opportunity to ask many questions tomorrow.

Email update and optional info sessions (pasting email)

6 months ago

Note: please let Nick and Satya know if you did not receive the email and optional invites for AI Crowd & Informa API sessions

Copy/pasting last week’s email update for those who might’ve missed:

Videos available here – please note that AI Crowd will re-record and Rodrigo/Aridhia could if necessary to be of better quality rather than the live versions:
Rodrigo’s workspace walkthrough available here:

AI Crowd submission walkthrough: video is available here:
note: Shivam just confirmed he will re-record a session:

There is now an additional example in python showing how to save, load, and submit a model – including adding a variable in both train and evaluate environments, see this post

Update on Hyderabad: As discussed – due to latency issues the core team brought up, Aridhia has volunteered to stand up a regional hub near Hyderabad. This has caused some days delay but for you could start – but these will be compensated for. Teams & workspaces have now been provisioned in Hyderabad. Check in on the Platform support channel for real time updates.

Is training data available during evaluation?

6 months ago

Hi all - quick update, we fully support making the training data available to the evaluation cluster. As Shivam mentioned - the training data will be visible to the evaluation cluster today. There were a few steps.

@yzhounvs - thanks for your concrete example of “data cleaning” - it is useful

I am traveling today but I would like to speak with you Monday if possible.

Raw data will be available soon but due to its size I was unable to find a quick solution while traveling.

New python example available

6 months ago

Dear all,

There is now an additional more detailed example in python showing how to save, load, and submit a model – including adding an invented variable in both train and evaluate environments.

There is the training notebook, the loading and predicting notebook, and it’s corresponding python script which could be used in submission.

you may find it on the Aridhia platform in:
/shared_data/software/Python/python_starter_kits

and on the main git lab:
Python_Starter_Kit_Training_pipeline.ipynb
Python_Starter_Kit_Prediction_pipeline.ipynb
predict_loading_model_example.py

Is training data available during evaluation?

6 months ago

But yes, for starters, I see no issue making the training data available in the evaluation cluster, and we will do so asap.

Regarding test data - I have discussed with the core team and will reach out directly with feedback and questions.

Informa API access and guide now available

6 months ago

The Informa API has been added to our proxy whitelist and should be accessible from all workspaces.

As announced, this will allow API access to the data and data models - to facilitate linking other data and directly extracting additional information from the trial database and drug database, including new data in the form of the events and catalysts database. Overview and details from Informa here

Please find the relevant materials for getting started with the API on the Aridhia platform:

Is training data available during evaluation?

6 months ago

Hi Shivam,
I will follow up with you regarding which paths the evaluation cluster should have access to.

Columns missing in the test dataset?

6 months ago

thanks for the quick work there shivam.

Notes from Monday's AI Crowd Q&A Session

6 months ago

=============================

  • divergence between data dictionary, and the data set

  • steppping through setting up the workspace/Rstudio and accessing the data would be helpful

  • test_release_small_dataset

  • Difference between Aridhia Team, DSAI team

  • Just basic things like operating on the data set in the workspace; use of SSH and VNC

  • whitelisting of sites (github, npm)

  • workspace allocation

  • instructions on use of git-lfs

  • teams on leaderboard

  • What happens if the solution is in R, with regards to anaconda

  • I cannot even access to the data in the shared_data folder. It is locked.

  • How to add team members to the repository

  • Copy Paste inside and outside the workspace : Chrome <> from outside the workspace

  • explain git tags

  • build docker images locally

    • Added in FAQ by shivam already
  • Conda get started instructions

    • install conda with wget + bash
    • create the first R-compatibe environment using conda create -f environment.yml (we should include this in the starter kit)
    • install.R in the starter kit
  • SSH Keys

    • to provide SSH keys, generate them
  • Install anaconda

  • AIcrowd FAQ : https://discourse.aicrowd.com/c/faq

  • Mohanty to fix bug of env var (in connors example

  • no linMod.rda (to be added by Connor)
    ####################################################################
    ####################################################################

  • Problem Statement (training, testing)

    • predict the probability of approval
  • Workspace (Aridhia)

  • Evaluation of submission

    • Software Runtime
    • Anaconda
      • Env Export

    Prediction
    - outcome columns not available in the testing data
    - probability of outcome

    Prediction
    - Structure of prediction file
    -

Example Prediction CSV

row_id, prob_approval
0,0.9
1,0.01
2,0.5



Gitlab submission

  • git clone git@gitlab.aicrowd.com:novartis/novartis-dsai-challenge-starter-kit.git

  • git remote remove origin

  • git remote add origin git@gitlab.aicrowd.com:mohanty/novartis-dsai-challenge-starter-kit-01.git

  • git tag -am “submission-v0.1” submission-v0.1

Example script shows how to create additional variables

6 months ago

Yes - I have an example and it will be available shortly - if not already. If I understand you correctly. I will check on this right now and include in the email which is set to be sent out today.

Endphaseyear in train and test set

6 months ago

Hi,

Q1: The split was done on OutcomeYear, which was removed from the training data set, as this would not be available directly following a phase 2.

EndphaseYear would be available, and was left in to carry the temporal signal, as approval base percentage has been declining.

Q2: Great question. All records are real trials. The trial ID’s are also as given in the raw datra and were not randomized.

Any tips on how to avoid Aridhia crashes / error 519 restarts?

6 months ago

@rodrigobarnes - would you like to comment on this one or reach out, or should we redirect directly to the MS team chat channel instead?

Intphaseendyear 1900

6 months ago

Thanks for the specifics. I have asked the informa team and am awaiting their response. I assume this is a known issue, and will relay.

Please consider releasing the datasets and use the submission file for evaluation

6 months ago

Regarding access, I will send a communication today. Due to latency issues in Hyderabad, another hub was stood up by Aridhia. The loss of time will be compensated for where possible.

Please consider releasing the datasets and use the submission file for evaluation

6 months ago

Hi and thanks for your suggestions.

  1. Will be discussed with the core team regarding data restrictions.

  2. a) If you need a GPU that can be arranged. Since one does not automatically need a GPU for deep learning - as Tensor Flow and similar able to be run on a CPU - we defaulted to CPU’s. I could easily imagine a GPU might prove necessary for considering new data types of larger sizes.
    b) We have a team who will be uploading all of our chemical assay data, linked to the core data set where possible - right now for phase 2. This required both legal approval and “wrangling”. We can certainly provide a linkage table for the compounds in the test set.

  3. You are not required to use the API. This was provided out of goodwill by Informa on a trial basis for the challenge. Many teams have data engineers and wranglers who have displayed quite some interest, and there will be a walk-through. It was also the proposed solution from Informa to get the linkage for the compound data. See next point regarding raw data.

  4. I will confirm with the core team, but you do or should shortly have access to the raw data for which we have a subscription. You can choose whether to participate in the leaderboard aspect – if so, it is your responsibility not to leak information if going back to the raw data. You can also simply proceed to the insights section and ignore the leaderboard. The leaderboard is a methods comparison, where additional data will be available centrally to all participants including the evaluation cluster where you can link in the same manner you did with the training data. The test split is meant to mimic post-phase 2 investment and conservatively avoid information leak. For practical reasons we chose a single date. Other options are rolling window, or re-training a model for each prediction based on all information time stamped earlier than the decision point, which was not feasible given the specifics of the data. Feel free to do as you feel appropriate for the insights section, including phase 3 design, etc

What's the meaning of "outcome"?

6 months ago

This is the outcome of the regulatory decision on approval.

There is a separate variable in the raw data called “Dev Status”.

Dev Status is our label. We treat development status launched and registered as successes or 1; discontinued, suspended, and no development reported as failures or 0; and phase ii clinical trial, phase iii clinical trial, and pre-registration as pipeline programs (i.e., still under development) which have no label.

Raw Informa data - availability where?

6 months ago

Thanks Bjoern, I will confirm in the morning with the Hyderabad team.

Intphaseendyear 1900

6 months ago

Hard to say without seeing the entire row.

While there was significant cleaning done on the data, at some point self-reporting can still have mistakes. And might require further filtering rules depending on team approach.

This might be something that could be cross-checked quickly. I’d be curious to know the results.

Which row number was it?

What should we submit in the end (besides the model)?

7 months ago

Hi Tianmeng -

What you submit as a presentation is completely up to you, but a Jupyter Notebook would be completely fine for example. And many teams will upload a powerpoint I’m sure.

For teams with a domain expert - they can help guide how to communicate and what sorts of insights would be most valuable to business decision makers for example. But there is no template. And furthermore, the evaluation committee will include a variety of functions from data science to portfolio optimization. In the coming weeks we will try to arrange a Q&A session with business problem owners on this topic.

Best, Nick

Workspace provisioning for teams, week of Nov 4th

7 months ago

Dear all,

Depending on whether your team was registered by you or assigned, your work space either has been or is being created by the Aridhia team. As detailed earlier, you will receive a confirmation when this process is over.

For any questions, reply below or email ned at Nediljko.Radanovic@aridhia.com

-Nick

Welcome to the Novartis DSAI Challenge Forums

7 months ago

Thank you so much for registering for the upcoming Data Science & AI challenge on clinical trial and program Probability of Success!

Please join our Challenge Walkthrough and Kick-off on Nov 11th. Here we will introduce you to:

  • The Core data set
  • Starter kits in python and/or R to load data, train, and submit predictions
  • The key challenges and insights teams should provide

In the meantime, please register your Azure account with our partner Aridhia:
https://workspaces.eastus.novartis.aridhia.io/ (US)
https://workspaces.northeurope.novartis.aridhia.io/ (other)

kelleni2 has not provided any information yet.