Q&A with Vishal Srivastava, Vice President (Model Validation) at Citi

Media 7 | September 8, 2021

Vishal Srivastava, Vice President (Model Validation) at Citi was invited as a keynote speaker to present on Fraud Analytics using Machine Learning at the International Automation in Banking Summit in New York in November 2019. Vishal has experience in quantitative risk modeling using advanced engineering, statistical, and machine learning technologies. His academic qualifications in combination with a Ph.D. in Chemical Engineering and an MBA in Finance have enabled him to challenge quantitative risk models with scientific rigor. Vishal’s doctoral thesis included the development of statistical and machine learning-based risk models—some of which are currently being used commercially. Vishal has 120+ peer-reviewed citations in areas such as risk management, quantitative modeling, machine learning, and predictive analytics.

As workplaces start to open, a hybrid model—seems to be a new norm that provides flexibility for people to operate both from their homes and offices, as we emerge out of the pandemic period.



MEDIA 7: Could you please tell us a little bit about yourself and what made you choose this career path?
VISHAL SRIVASTAVA:
Since my childhood, I have had a deep interest in math and science—which led me to pursue a bachelor’s in engineering degree at NIT Trichy (National Institute of Technology, Tiruchirappalli) in India. Later, to advance my knowledge, I pursued MBA and Ph.D. studies in the United States, with fully-funded university scholarships. During my Ph.D. research, I was intrigued by various applications of mathematics with which risk in engineering systems could be quantified. Thanks to my advisors Prof. Carolyn Koh and Prof. Luis Zerpa at the Colorado School of Mines, I got the opportunity to explore ideas—from first principles to machine learning—and to build risk modeling frameworks in high-pressure flow systems. As a result of my Ph.D., we were able to develop risk frameworks to be used by consortium partners that included leading global energy companies. My Ph.D. research in the quantification of risk was an intellectually stimulating experience that taught me that anything is possible if we let our focus and energy stick to a single idea over a reasonable time.

Due to the nature of my Ph.D. research—which included quantitative risk modeling—and my earlier degree of MBA in Finance, I was contacted by several risk management professionals for a potential job opportunity in the finance sector. From a mathematical standpoint, risk management in engineering and finance has a lot of overlap.  The computation of risks in engineering systems deals with investigating factors that can lead to a system failure, which can be predicted using first principles-based engineering methods or with statistical models that include the historical distribution of failure events. Similarly, credit risk management can be approached using the first principle-based mathematical methods or statistical models that forecast defaults as a function of macroeconomic or account level variables. In both cases, a binary classification model can be developed in modeling these default or failure events. I found it fascinating to explore the different avenues where a graduate study in risk engineering could be applied.

About six months before my Ph.D. defense, I had an offer from Bank of the West, BNP Paribas in model risk division. My job role as an Assistant Vice President in the Model Risk Team was to challenge the model-building process of credit and fraud risk models—both involved binary classification models. The credit risk models include a logistic regression framework which is a well-accepted industry methodology for classification and is easy to interpret. The fraud risk models included both—traditional rule-based models—and new age RNN (Recurrent Neural Network) based sequential models, which use complex and non-linear models. From this experience, I learned that from a regulatory standpoint—model explainability could be a key factor while selecting a model. This was a valuable experience, but I’ve always enjoyed challenging myself and moving out of my comfort zone. So, about one and half years later, I accepted an opportunity to work as Vice President with Citibank’s Model Risk Division in the Secured Loan team, where my responsibilities included working with the international model validation team-members to review International and US Mortgage default risk models. My focus at this job is to challenge mortgage default risk models across various continents to ensure that these models are regulatory compliant. This experience is extremely insightful due to the varied nature of credit default events across different continents as well as the homogeneity in the modeling approach towards developing a model.


M7: What are some of the means through which you select appropriate model validation methodology?
VS:
In an increasingly competitive environment, financial institutions depend on models which help them optimize risks and make decisions that are well informed. Model validation managers need to ensure that every step in the model building process—data acquisition, conceptual soundness evaluation, model stability analysis, back-testing, performance assessment, model implementation testing—is well supported by a sound scientific framework. This is done to ensure that critical decisions such as loss estimates, capital allocation, and budget planning are taken based on scientific and mathematical reasoning rather than intuition. One key aspect in the whole model validation process is to ensure that the given model is compliant with the prevailing regulatory framework. In that regard, model developers present an assessment of all model usages and outputs. The performance assessment is conducted for all model usages and model outputs across all forecasting horizons. But one caveat of this process is that model risk assessment across all models can be cost-intensive. Therefore, the model review process is prioritized and models of higher importance—that are of substantial size and with significant risk contribution—are reviewed with a greater frequency. These are some of the key guidelines model validators keep in mind while performing model risk management activities.


The US economy has stayed resilient for most of 2021 when macroeconomic factors such as consumer spending and the unemployment rate have been showing promising trends.



M7: What are some of your go-to model validation techniques that help you effectively identify and manage model risk?
VS:
There can be no fixed technique that can be homogeneously applied to evaluate if a model under review is totally fit for the purpose. However, at a high level, there can be some guiding principles that could be quite useful while deciding to approve or reject a model. The first check is to ensure if there has been enough analysis performed on the conceptual soundness of the final selected methodology which is proposed in the model. Here the goal is to ensure that there is sufficient evidence to justify if the selected methodology is indeed the right modeling approach. For example, for the scorecard model, one can use logistic regression, decision tree, or neural network model. In such a situation, the model validator would review if enough analysis has been performed to justify if the given modeling framework suits the given data best and if the selected model can be sufficiently explained to the regulators. 

Additionally, model developers also explore the alternative modeling framework to demonstrate why the selected modeling framework is superior to the alternative modeling methodologies. The next aspect in model validation is to find if there are any inadequacies towards analysis or model documentation. If that is observed during the validation, the same needs to be recorded in the model validation report as findings and recommendations. In the model validation report, the model validator provides a record of comprehensive documentation to record all model findings and recommendations. This serves later as a reference document for model developers when there is a need for future model enhancement. Next, model developers need to ensure if model assumptions continue to be reasonable and are based on sound theoretical appropriateness. Consequences of model assumptions violations can be expensive. As an example, during the financial crisis of 2007–2008, several modelers assumed that the housing market will continue to grow based on the historical performance and previous data. However, during the financial crisis of 2007–2008, the housing market plunged, and many assumptions of those times were violated. As a result, several companies had to face a huge financial loss. Hence, it is imperative that each of the model assumptions is carefully evaluated. Model validators also need to ensure if the data quality checks have been performed sufficiently. The goal here is to ensure a scientific approach towards data segmentation, data cleaning, data sampling methodology, missing values, and data outliers—which can severely affect the model forecasts. The model validator also needs to ensure if data sources—both internal and external (rating agencies, etc.) are well checked and properly recorded while clearly justifying all data exclusions. The model validator also needs to ensure if the model developer has performed a sound variable selection process and if all variable transformations are well documented. Many times, continuous variables are converted to a categorical variable by a process called binning, and dummy variables are created. Any discrepancy in the variable transformation in the modeling and implementation stage can lead to a big discrepancy between the modeling and production. Another very important part of the model validation exercise is model back-testing and performance analysis.  This is to ensure that model is still producing accurate forecasts even for the recent period with unseen data. As described, the three main pillars of the model validation process can be depicted as below:
 

 



Model validation managers need to ensure that every step in the model building process—data acquisition, conceptual soundness evaluation, model stability analysis, back-testing, performance assessment, model implementation testing—is well supported by a sound scientific framework.

Model validator reviews if the model developer has performed back-testing in OOT (out-of-time) and OOS (out-of-sample) data to ascertain if the model is still accurate when the sample is not from the data that was used in the original developmental period to rule out overfitting. Next, the model validator must ensure if the model is meeting all the necessary regulatory compliance and all the model document fully complies with the necessary regulatory requirements. Model validators also need to review model dependencies. For instance, if the output from one model goes as an input to the second model, and if there is a performance issue with the first model, the performance of the second model can be adversely affected due to model dependencies. These are some of the pointers that model validators use to review a given model. A summary of the model validation review process can be pictorially represented in the below diagram:


M7: What do you see as the most noticeable change right now happening in the workforce, encouraged by the rise of digital technologies?
VS:
There is a Chinese proverb that says— “May you live in interesting times”. If we look around, we are rather living currently in transformational times that will redefine our future. Many banking tasks which earlier required physical proximity, are now being automated with digital innovations—that include advancements in computer vision and image recognition. Financial institutions have already introduced several innovative products—from automatic cheque deposits and online cash transfers to digital payments and transactions.  Additionally, the rise of digital technologies coupled with the changes due to the pandemic has brought irreversible changes in our workforce. As workplaces start to open, a hybrid model—seems to be a new norm that provides flexibility for people to operate both from their homes and offices, as we emerge out of the pandemic period.  There is an immense opportunity to retain the best parts of office culture while getting freedom from inefficient tasks and office meetings, which are unproductive.  This is resulting in the trend that commercial workplaces are moving into residential complexes as organizations are exploring new opportunities to be more efficient. We are seeing a new form of organizational agility, which is empowering teamwork across all disciplines and offshore locations. In my opinion, companies that quickly adapt to this remotely operated flexi-time organizational culture—rather than enforce the orthodoxy of 9-to-5 office-centric work—will have a clear competitive advantage in this new era of work. As digital transactions take precedence, many banking products such as payments and other forms of deposits—are fast becoming obsolete because people are able to use these applications on their cell phones. The ongoing pandemic has accelerated the adoption of automation and AI processes which were started in the pre-covid period. All these changes create immense opportunities in the financial sector in general.


M7: What are the top challenges you see for the industry in general?
VS:
The year 2021 is full of changes in many aspects. First, due to the rapid increase in pandemic cases worldwide, many countries witnessed some sort of slow-down in their economy during last year. However, with the ongoing vaccination drive, and reopening of offices and workplaces, synchronous global recovery has also been witnessed in the recent period. The US economy has stayed resilient for most of 2021 when macroeconomic factors such as consumer spending and the unemployment rate have been showing promising trends. However, the unemployment rate in the US for last year was among the highest in the last several decades. The dynamics and volatility in macroeconomic drivers thus affected many modeling forecasts. This is one of the main challenges from a model risk standpoint when many traditional models don’t seem to work as well as they did during the pre-pandemic time. The rise in macroeconomic volatility in the wake of COVID-19 has increased the uncertainty in modeling forecasts. When this uncertainty is not handled in a sound manner, this could result in two things—An inaccurate forecast from a simple model or a need towards a more complex model, giving rise to overfitting problems. From a model risk validation standpoint, model complexity is a growing challenge in the current times as many products are seeing the adoption of AI and machine learning to make the best use of banking data for improving efficiencies and gaining competitive intelligence. For such models, there is a need for modelers to explain the working of the model not just the performance of the model. With greater use of AI and analytics in the model risk domain, model explainability becomes a challenge faced by modelers. However, there have been significant advancements in model interpretability aspects with Explainable AI due to techniques such as LIME (Local Interpretable Model-agnostic Explanations) and SHAP, which stands for SHapely Additive exPlanation. Nevertheless, it is a constant battle to strike the right balance between model accuracy and model explainability in the wake of regulatory requirements. From a compliance viewpoint, this could also result in an environment that requires greater regulatory intervention in the model risk domain. These are some of the main challenges faced in the model risk domain from a technical standpoint. From a human resource viewpoint, finding good talent in the model risk domain is a big challenge in current times when many technology companies are hiring data scientists for similar roles. All challenges however come with great opportunities. Financial institutions are innovating and offering products that are creative and user-friendly. The speed of innovation has improved and the future only looks more promising.


M7: When you are not working, what else are you seen doing?
VS:
I love jogging and hiking in nature. I have recently finished a 100-day challenge of jogging 3 miles a day without missing a single day and I hope to take this to a next level by joining a marathon in Dallas when I move there next week. Apart from that, I love listening to podcasts on a variety of subjects. I have been recently listening to podcasts of Rich Rolls and Andrew Huberman, a neuroscientist from Stanford, who publicly presents his research about neuroscience and all the fun experiments his team performs at Stanford University. I also enjoy exploring different types of meditations and like to read about the healing effects of meditation. Other than these, I also enjoy swimming and vacationing to hilly places.

��


ABOUT CITIBANK

Citibank is one of the leading financial institutions of the world and is headquartered in New York City. It has one of the largest customer bases and has served more than 200 million clients with operations in more than 160 countries in the world. The U.S. branches are concentrated in six metropolitan areas: New York, Chicago, Los Angeles, San Francisco, Washington, D.C., and Miami. In addition, Citi is also a leading philanthropist company that is focused on catalyzing sustainable growth through transparency, innovation, and market-based solutions.

More THOUGHT LEADERS

'Raising the voices of those who may not always be heard is critical,' says Claire Thomas

Media 7 | April 28, 2023

Claire Thomas is responsible for developing and implementing a strategy for diversity, equity, and inclusion (DEI) across Hitachi Vantara through programs that reflect the diverse backgrounds, interests, and passions of their current and future workforce. Continue reading to learn her views on the significance of inclusion and diversity in an organization....

Read More

Q&A with Charles Southwood, Vice President, N. Europe and MEA at Denodo

Media 7 | September 15, 2021

Charles Southwood, Regional VP at Denodo Technologies is responsible for the company’s business revenues in Northern Europe, Middle East and South Africa. He is passionate about working in rapidly moving and innovative markets to support customer success and to align IT solutions that meet the changing business needs. With a degree in engineering from Imperial College London, Charles has over 20 years of experience in data integration, big data, IT infrastructure/IT operations and Business Analytics....

Read More

Q&A with Sadiqah Musa, Co-Founder at Black In Data

Media 7 | September 1, 2021

Sadiqah Musa, Co-Founder at Black In Data, is also an experienced Senior Data Analyst at Guardian News and Media with a demonstrated history of working in the energy and publishing sectors. She is skilled in Advanced Excel, SQL, Python, data visualization, project management, and Data Analysis and has a strong professional background with a Master of Science (MSc) from The University of Manchester....

Read More

'Raising the voices of those who may not always be heard is critical,' says Claire Thomas

Media 7 | April 28, 2023

Claire Thomas is responsible for developing and implementing a strategy for diversity, equity, and inclusion (DEI) across Hitachi Vantara through programs that reflect the diverse backgrounds, interests, and passions of their current and future workforce. Continue reading to learn her views on the significance of inclusion and diversity in an organization....

Read More

Q&A with Charles Southwood, Vice President, N. Europe and MEA at Denodo

Media 7 | September 15, 2021

Charles Southwood, Regional VP at Denodo Technologies is responsible for the company’s business revenues in Northern Europe, Middle East and South Africa. He is passionate about working in rapidly moving and innovative markets to support customer success and to align IT solutions that meet the changing business needs. With a degree in engineering from Imperial College London, Charles has over 20 years of experience in data integration, big data, IT infrastructure/IT operations and Business Analytics....

Read More

Q&A with Sadiqah Musa, Co-Founder at Black In Data

Media 7 | September 1, 2021

Sadiqah Musa, Co-Founder at Black In Data, is also an experienced Senior Data Analyst at Guardian News and Media with a demonstrated history of working in the energy and publishing sectors. She is skilled in Advanced Excel, SQL, Python, data visualization, project management, and Data Analysis and has a strong professional background with a Master of Science (MSc) from The University of Manchester....

Read More

Related News

BUSINESS INTELLIGENCE, BIG DATA MANAGEMENT, DATA SCIENCE

NorthGravity Launches Hubcio: Revolutionizing Natural Language Generation for Data Insights as Part of the NorthGravity AI Suite

Newswire | May 29, 2023

NorthGravity, a leading provider of advanced data analytics software, is thrilled to announce the release of Hubcio. With its cutting-edge capabilities, Hubcio will join the NorthGravity AI (NGAI), sitting alongside the NorthGravity Platform (NGP) and the NorthGravity Insights (NGI). Hubcio's integration with the NorthGravity AI Suite enables users to seamlessly leverage its power. Users can effortlessly build advanced data pipelines, streamline Python code to extract valuable insights, and gain a comprehensive view of internal documents and data—all within the unified NGAI ecosystem. "We are excited to announce that Hubcio will be a vital component of our NorthGravity product," said Travis Nadelhoffer, CEO of NorthGravity. "By integrating Hubcio with NGP and NGI, we are providing our clients with a powerful suite of AI tools that seamlessly work together to enable data-driven insights and decision-making." Key features of Hubcio within the NGAI include: Unified AI Ecosystem: Hubcio seamlessly integrates with the NorthGravity Platform (NGP) and NorthGravity Insights (NGI) as part of the NorthGravity AI Suite (NGAI), providing users with an integrated environment for data analysis. Advanced Data Pipelines: Users can build sophisticated data pipelines utilizing the NGAI ecosystem within the NGP, effortlessly transforming and processing data to extract valuable insights. Python Optimization: Hubcio simplifies the process of writing Python code, enabling users to extract maximum value from data with minimal effort. Advanced Expeditor: Hubcio creates a foundation for building data pipelines and writing code within the NGP. Internal Document Insights: Gain a comprehensive view of internal documents and data, uncovering valuable insights and knowledge hidden within your organization, powered by the NGAI ecosystem. Automated Submission and Training: Automate the submission and training of internal data and documents, creating a secure and customizable knowledge base for natural language queries, all within the NG ecosystem. To learn more about Hubcio, please visit https://www.northgravity.com/. About NorthGravity NorthGravity is a leading provider of advanced data analytics software, helping businesses unlock the full potential of their data. NorthGravity's innovative software suite includes the NorthGravity Platform (NGP), NorthGravity Artificial Intelligence (NGAI), and NorthGravity Insights (NGI). The NGP is a platform used for data storage, collection, and pipelines. The NGAI is a suite of AI tasks, autoML and Hubcio. The NGI is a business insights tool. NorthGravity is committed to delivering superior software that drives business growth and success.

Read More

BIG DATA MANAGEMENT, DATA SCIENCE

HighByte Releases New API Gateway to OT Systems, Unlocking Industrial Data for the Enterprise

Prnewswire | May 30, 2023

HighByte®, an industrial software company, today announced the release of HighByte Intelligence Hub version 3.1 that expands operational technology (OT) data access, governance, and scale capabilities for industrial companies. The release includes a new REST Data Server that acts as an API gateway for industrial data residing in OT systems, so any application or service with an HTTP client can securely request OT data in raw or modeled form directly from the Intelligence Hub—without requiring domain knowledge of the underlying systems. "With its initial release in 2020, HighByte Intelligence Hub entered the market as a client-based application capable of publishing data to consuming systems. Earlier this year, we released server-based capabilities with the addition of an embedded MQTT broker, allowing systems to subscribe to industrial data sources. In a few short months, we have expanded data access yet again with the ability to request data on-demand from the Intelligence Hub through a REST API," said HighByte CEO Tony Paine. "This technical evolution reflects the maturing Industrial DataOps market and the increasingly sophisticated use cases of our customers. We are listening and committed to staying ahead of market needs." HighByte Intelligence Hub version 3.1 adds additional capabilities to support data governance and scale, including more flexible configuration management. The Intelligence Hub now allows easy export and import of full or partial project configurations, which can be configured through the user interface, configuration files, or the configuration API. Users can also import models and instances directly from third party systems like Element Unify, supporting AWS's Industrial Data Fabric architecture. In addition, users can dynamically drive template definitions from third party sources and trigger flows for individual assets, streamlining configuration across similar assets and use cases. "We are particularly excited about the new ability to import and export full or partial projects directly through the user interface. This functionality holds special significance for us as we are running HighByte Intelligence Hub in multiple Docker containers," said Matthew Venter, Solutions Architect at Gousto, an online meal-kit manufacturer and retailer and certified B Corporation. "With the ability to seamlessly import and export projects, we can effortlessly replicate and distribute our configurations across different container instances." Other features include a new Apache Kafka connector and significant improvements to existing connectors including PI System, Sparkplug, OPC UA, REST Client, AWS IoT SiteWise, Modbus, and the File connector (now with FTP support). HighByte Intelligence Hub version 3.1 is now commercially available. All new features and capabilities introduced in version 3.1 are included in standard pricing. Please contact HighByte or an authorized distributor to request a trial or purchase an annual subscription license. About HighByte HighByte is an industrial software company in Portland, Maine USA building solutions that address the data architecture and integration challenges created by Industry 4.0. HighByte Intelligence Hub, the company's award-winning Industrial DataOps software, provides modeled, ready-to-use data to the Cloud using a codeless interface to speed integration time and accelerate analytics. Learn more at https://www.highbyte.com.

Read More

BIG DATA MANAGEMENT, DATA SCIENCE, MACHINE LEARNING

Snorkel AI Announces Third Annual Future of Data-Centric AI Conference

Businesswire | May 30, 2023

Snorkel AI, the data-centric AI platform company, today announced the agenda for Future of Data-Centric AI Conference 2023 (FDCAI). Now in its third year, FDCAI has established itself as the definitive community and conference for AI, bringing together thousands of data scientists, ML engineers, software developers, Fortune 500 organizations, AI-first companies, and academia under one roof to share best practices and the latest in AI innovation. Held on June 7-8, 2023, this year's virtual event will explore how to leverage new AI advances such as large language models (LLMs), foundation models, generative AI, programmatic labeling, weak supervision, prompting, synthetic data and more—all using data-centric AI workflows. "AI has seen an explosion of innovation in the last few months prompted by the fast rise of large language models,” said Devang Sachdev, VP of Marketing, Snorkel AI. “These advancements, while remarkable, also underscore the urgent need for adopting data-centric approaches to ensure high-quality data and effective AI development. This is why we started The Future of Data-Centric AI conference three years ago, and why we’re continuing to build on it today—to create a forum for the world’s AI innovators to discuss how to deploy AI in real-world settings.” FDCAI’s speaker line-up this year includes leading AI practitioners, researchers and academics from such companies as: Arista, Bank of America, Bloomberg, Capital One, Caterpillar, Comcast, Databricks, EY, Georgetown University CSET, Google, Harvard, Hugging Face, JP Morgan & Chase, Kaiser Permanente, Marsh, Mayo Clinic, Microsoft, Natwest and many others. Notable mentions include: DJ Patil, first U.S. Chief Data Scientist and General Partner, Greatpoint Ventures Emad Mostaque, Founder and CEO, Stability AI Yoav Shoham, Co-founder, AI21 Labs Matei Zaharia, Chief Technologist, Databricks Nurtekin Savas, Head of Enterprise Data Science, Capital One Gideon Mann, Head of ML Product and Research, Office of the CTO at Bloomberg LP. The sessions at this year’s conference will focus on the following areas: Data development techniques: programmatic labeling, active learning, weak supervision, data cleaning, and synthetic data augmentation. Enterprise use cases: predictive AI, generative AI, NLP, computer vision, conversational AI. AI development stack: AutoML, ML frameworks, no-code/low-code development. Foundation models/LLMs: fine-tuning, prompt engineering, prompt chaining, and enterprise adoption This year’s conference also features an AI poster competition with $15K in prizes and submissions from Argonne National Lab, Columbia University, Cornell University, Medidata, Stanford University, TitanML, University of Cambridge, University of Toronto, University of Utah and University of Wisconsin-Madison. The Future of Data-Centric AI conference is hosted by Snorkel AI in partnership with Gretel AI, Hugging Face, Lambda Labs, Microsoft, Predibase, Seldon AI, and Together. All the sessions are virtual and free to attend. For the full conference schedule or to register, visit: https://future.snorkel.ai/ About Snorkel AI Founded by a team spun out of the Stanford AI Lab, Snorkel AI makes AI application development fast and practical by unlocking the power of machine learning without the bottleneck of manually-labeled training data. Snorkel Flow is the first data-centric AI platform powered by programmatic labeling. Backed by Addition, Greylock, GV, In-Q-Tel, Lightspeed Venture Partners and funds and accounts managed by BlackRock, the company is based in Palo Alto. For more information on Snorkel AI, please visit: https://www.snorkel.ai/ or follow @SnorkelAI.

Read More

BUSINESS INTELLIGENCE, BIG DATA MANAGEMENT, DATA SCIENCE

NorthGravity Launches Hubcio: Revolutionizing Natural Language Generation for Data Insights as Part of the NorthGravity AI Suite

Newswire | May 29, 2023

NorthGravity, a leading provider of advanced data analytics software, is thrilled to announce the release of Hubcio. With its cutting-edge capabilities, Hubcio will join the NorthGravity AI (NGAI), sitting alongside the NorthGravity Platform (NGP) and the NorthGravity Insights (NGI). Hubcio's integration with the NorthGravity AI Suite enables users to seamlessly leverage its power. Users can effortlessly build advanced data pipelines, streamline Python code to extract valuable insights, and gain a comprehensive view of internal documents and data—all within the unified NGAI ecosystem. "We are excited to announce that Hubcio will be a vital component of our NorthGravity product," said Travis Nadelhoffer, CEO of NorthGravity. "By integrating Hubcio with NGP and NGI, we are providing our clients with a powerful suite of AI tools that seamlessly work together to enable data-driven insights and decision-making." Key features of Hubcio within the NGAI include: Unified AI Ecosystem: Hubcio seamlessly integrates with the NorthGravity Platform (NGP) and NorthGravity Insights (NGI) as part of the NorthGravity AI Suite (NGAI), providing users with an integrated environment for data analysis. Advanced Data Pipelines: Users can build sophisticated data pipelines utilizing the NGAI ecosystem within the NGP, effortlessly transforming and processing data to extract valuable insights. Python Optimization: Hubcio simplifies the process of writing Python code, enabling users to extract maximum value from data with minimal effort. Advanced Expeditor: Hubcio creates a foundation for building data pipelines and writing code within the NGP. Internal Document Insights: Gain a comprehensive view of internal documents and data, uncovering valuable insights and knowledge hidden within your organization, powered by the NGAI ecosystem. Automated Submission and Training: Automate the submission and training of internal data and documents, creating a secure and customizable knowledge base for natural language queries, all within the NG ecosystem. To learn more about Hubcio, please visit https://www.northgravity.com/. About NorthGravity NorthGravity is a leading provider of advanced data analytics software, helping businesses unlock the full potential of their data. NorthGravity's innovative software suite includes the NorthGravity Platform (NGP), NorthGravity Artificial Intelligence (NGAI), and NorthGravity Insights (NGI). The NGP is a platform used for data storage, collection, and pipelines. The NGAI is a suite of AI tasks, autoML and Hubcio. The NGI is a business insights tool. NorthGravity is committed to delivering superior software that drives business growth and success.

Read More

BIG DATA MANAGEMENT, DATA SCIENCE

HighByte Releases New API Gateway to OT Systems, Unlocking Industrial Data for the Enterprise

Prnewswire | May 30, 2023

HighByte®, an industrial software company, today announced the release of HighByte Intelligence Hub version 3.1 that expands operational technology (OT) data access, governance, and scale capabilities for industrial companies. The release includes a new REST Data Server that acts as an API gateway for industrial data residing in OT systems, so any application or service with an HTTP client can securely request OT data in raw or modeled form directly from the Intelligence Hub—without requiring domain knowledge of the underlying systems. "With its initial release in 2020, HighByte Intelligence Hub entered the market as a client-based application capable of publishing data to consuming systems. Earlier this year, we released server-based capabilities with the addition of an embedded MQTT broker, allowing systems to subscribe to industrial data sources. In a few short months, we have expanded data access yet again with the ability to request data on-demand from the Intelligence Hub through a REST API," said HighByte CEO Tony Paine. "This technical evolution reflects the maturing Industrial DataOps market and the increasingly sophisticated use cases of our customers. We are listening and committed to staying ahead of market needs." HighByte Intelligence Hub version 3.1 adds additional capabilities to support data governance and scale, including more flexible configuration management. The Intelligence Hub now allows easy export and import of full or partial project configurations, which can be configured through the user interface, configuration files, or the configuration API. Users can also import models and instances directly from third party systems like Element Unify, supporting AWS's Industrial Data Fabric architecture. In addition, users can dynamically drive template definitions from third party sources and trigger flows for individual assets, streamlining configuration across similar assets and use cases. "We are particularly excited about the new ability to import and export full or partial projects directly through the user interface. This functionality holds special significance for us as we are running HighByte Intelligence Hub in multiple Docker containers," said Matthew Venter, Solutions Architect at Gousto, an online meal-kit manufacturer and retailer and certified B Corporation. "With the ability to seamlessly import and export projects, we can effortlessly replicate and distribute our configurations across different container instances." Other features include a new Apache Kafka connector and significant improvements to existing connectors including PI System, Sparkplug, OPC UA, REST Client, AWS IoT SiteWise, Modbus, and the File connector (now with FTP support). HighByte Intelligence Hub version 3.1 is now commercially available. All new features and capabilities introduced in version 3.1 are included in standard pricing. Please contact HighByte or an authorized distributor to request a trial or purchase an annual subscription license. About HighByte HighByte is an industrial software company in Portland, Maine USA building solutions that address the data architecture and integration challenges created by Industry 4.0. HighByte Intelligence Hub, the company's award-winning Industrial DataOps software, provides modeled, ready-to-use data to the Cloud using a codeless interface to speed integration time and accelerate analytics. Learn more at https://www.highbyte.com.

Read More

BIG DATA MANAGEMENT, DATA SCIENCE, MACHINE LEARNING

Snorkel AI Announces Third Annual Future of Data-Centric AI Conference

Businesswire | May 30, 2023

Snorkel AI, the data-centric AI platform company, today announced the agenda for Future of Data-Centric AI Conference 2023 (FDCAI). Now in its third year, FDCAI has established itself as the definitive community and conference for AI, bringing together thousands of data scientists, ML engineers, software developers, Fortune 500 organizations, AI-first companies, and academia under one roof to share best practices and the latest in AI innovation. Held on June 7-8, 2023, this year's virtual event will explore how to leverage new AI advances such as large language models (LLMs), foundation models, generative AI, programmatic labeling, weak supervision, prompting, synthetic data and more—all using data-centric AI workflows. "AI has seen an explosion of innovation in the last few months prompted by the fast rise of large language models,” said Devang Sachdev, VP of Marketing, Snorkel AI. “These advancements, while remarkable, also underscore the urgent need for adopting data-centric approaches to ensure high-quality data and effective AI development. This is why we started The Future of Data-Centric AI conference three years ago, and why we’re continuing to build on it today—to create a forum for the world’s AI innovators to discuss how to deploy AI in real-world settings.” FDCAI’s speaker line-up this year includes leading AI practitioners, researchers and academics from such companies as: Arista, Bank of America, Bloomberg, Capital One, Caterpillar, Comcast, Databricks, EY, Georgetown University CSET, Google, Harvard, Hugging Face, JP Morgan & Chase, Kaiser Permanente, Marsh, Mayo Clinic, Microsoft, Natwest and many others. Notable mentions include: DJ Patil, first U.S. Chief Data Scientist and General Partner, Greatpoint Ventures Emad Mostaque, Founder and CEO, Stability AI Yoav Shoham, Co-founder, AI21 Labs Matei Zaharia, Chief Technologist, Databricks Nurtekin Savas, Head of Enterprise Data Science, Capital One Gideon Mann, Head of ML Product and Research, Office of the CTO at Bloomberg LP. The sessions at this year’s conference will focus on the following areas: Data development techniques: programmatic labeling, active learning, weak supervision, data cleaning, and synthetic data augmentation. Enterprise use cases: predictive AI, generative AI, NLP, computer vision, conversational AI. AI development stack: AutoML, ML frameworks, no-code/low-code development. Foundation models/LLMs: fine-tuning, prompt engineering, prompt chaining, and enterprise adoption This year’s conference also features an AI poster competition with $15K in prizes and submissions from Argonne National Lab, Columbia University, Cornell University, Medidata, Stanford University, TitanML, University of Cambridge, University of Toronto, University of Utah and University of Wisconsin-Madison. The Future of Data-Centric AI conference is hosted by Snorkel AI in partnership with Gretel AI, Hugging Face, Lambda Labs, Microsoft, Predibase, Seldon AI, and Together. All the sessions are virtual and free to attend. For the full conference schedule or to register, visit: https://future.snorkel.ai/ About Snorkel AI Founded by a team spun out of the Stanford AI Lab, Snorkel AI makes AI application development fast and practical by unlocking the power of machine learning without the bottleneck of manually-labeled training data. Snorkel Flow is the first data-centric AI platform powered by programmatic labeling. Backed by Addition, Greylock, GV, In-Q-Tel, Lightspeed Venture Partners and funds and accounts managed by BlackRock, the company is based in Palo Alto. For more information on Snorkel AI, please visit: https://www.snorkel.ai/ or follow @SnorkelAI.

Read More

Spotlight

Events

Resources

resource image

BUSINESS INTELLIGENCE, BIG DATA MANAGEMENT

SAP Datashpere

Whitepaper

resource image

PERFORMANCE MANAGEMENT

The Aftermath of the Great Resignation

Article

resource image

BUSINESS INTELLIGENCE, BIG DATA MANAGEMENT

Are Predictive Analytics Truly Predictive?

Whitepaper

resource image

BUSINESS INTELLIGENCE, BIG DATA MANAGEMENT

SAP Datashpere

Whitepaper

resource image

PERFORMANCE MANAGEMENT

The Aftermath of the Great Resignation

Article

resource image

BUSINESS INTELLIGENCE, BIG DATA MANAGEMENT

Are Predictive Analytics Truly Predictive?

Whitepaper

Events