Data Scientist

  • Permanent
  • Anywhere

Website Bigbear.ai


Company : Bigbear.ai

Country: United States

Location : Chantilly, Virginia

Post: Sat, 04 Mar 2023 19:01:55 GMT

Expires: Sat, 01 Apr 2023 23:59:59 GMT

Apply Job : Apply Online

—————————— Job Description ——————————

Bigbear.ai is seeking an experienced Data Scientist to join our team to support an exciting opportunity as part of our Virtual Anticipation Network (VANE) and Dominate Federal Product efforts.  This position will be fully remote within the United States. 


This is an opportunity to be part of one of the fastest growing AI/ML companies in the industry. At BigBear.ai, we’re in this business together. We own it, we make it thrive, and we enjoy the challenges of our work. We know that our employees play the largest role in our continual success. That is why we foster an environment of growth and development, with an emphasis on opportunity, recognition, and work-life balance. We give the same high level of commitment to our employees that we give to our clients. If BigBear.ai sounds like the place where you want to be, we’d enjoy speaking with you.

What you will do:
Work as a part of a multi-disciplined team, including software engineers, data scientists, data engineers, and intelligence analysts to develop our VANE and Dominate federal product solutions.

Build and develop AI/ML models, including natural language processing and time series forecasting techniques, to be applied to structured and unstructured data sets.

Work to identify, evaluate, ingest, and integrate new sources of data. Where possible, advise the project team regarding opportunities to cleanse the data before ingestion.

Assess the quality and value of potential data sources. Periodically review the quality of existing data sources and provide recommendations for improvement.

Manufacture representative, notional data for use by the development team when representative data is not available.

Validate the outputs of the solution.

What you need to have:
Bachelor’s Degree and 8+ years of experience; Master’s Degree and 6+ years of experience; PhD and 3+ years of experience (in lieu of Bachelor’s degree, 6 additional years of relevant experience)

Clearance: Must be able to obtain and maintain an active Secret clearance. 

Experience of architecting solutions in a core discipline of Big Data, Machine Learning, or Data Analytics

An ability to communicate at an appropriate level complex technical concepts to both the development team.

An ability to work independently with large data sets and make judgements based on careful analysis.

A curiosity toward data analysis and the ability to answer questions the customer is not yet able to articulate.

A high-level of attention to detail.

An ability to work independently.

An ability to work with a cross-functional team in a highly collaborative environment and communicate feedback constructively.

Familiarity and ability to develop – or quickly ramp up to develop – with the following:ML Libraries such as Scikit-learn, Pandas, Numpy, PySpark, MXNet, MLib, Weka, Spacy, and FastText

ML Platforms such as KNIME, Spark, SageMaker, H2O, and TensorFlow

Java, Python, and SQL including related ecosystems and frameworks, like Eclipse, Jupyter Notebooks, Spring, and Djanjo

Distributed platforms (i.e. HBase, PrestoDB, Athena, Spark, Kafka) and the cloud (i.e. AWS, GCP, Azure)

Performing dimensionality reduction and feature selection

Automating the assessment of model validity to drive hyper-parameter tuning

Transforming data to create stationary inputs, especially when facing heteroskedasticity, periodicity, and trends

Interfacing with RESTful APIs

Extract-transform-load (ETL) data pipelines, especially using KNIME, Informatica, Talend, CloverDX, Apache NiFi, or other frameworks

Productionalizing experimental code, optimizing for compute levels and processing time

What we’d like you to have:
An ability to think critically. All members of the team have an opportunity to shape the future of the project. The ability to think critically and ask if the team’s current path serves the customer’s interests is crucial to delivering a quality product.

Demonstrated database experience with bonus points for Elasticsearch, S3, Athena, and KNIME.

Demonstrated experience working as part of a software team using Agile methods

Demonstrated experience with time series forecasting, including tensor completion and VAR.

Demonstrated experience with natural language processing techniques, including applying novel transformations and enhancements to unstructured data.

Demonstration experience with tuning algorithms and applying new methodologies to problem sets.

About BigBear.ai:
BigBear.ai delivers AI-powered analytics and cyber engineering solutions to support mission-critical operations and decision-making in complex, real-world environments. BigBear.ai’s customers, which include the US Intelligence Community, Department of Defense, the US Federal Government, as well as customers in manufacturing, healthcare, commercial space, and other sectors, rely on BigBear.ai’s solutions to see and shape their world through reliable, predictive insights and goal-oriented advice. Headquartered in Columbia, Maryland, BigBear.ai is a global, public company traded on the NYSE under the symbol BBAI. For more information, please visit:  and follow BigBear.ai on Twitter: .

To apply for this job please visit www.resume-library.com.