About SISO University

SISO University (SISO U) is a training program aimed at M&S practitioners, which offers various levels of M&S courses to SISO members. This program spans multiple SIWs (not every tutorial will be offered at every workshop) to help SISO members develop a persistent body of knowledge. The courses are split into three levels: 100 - Overview, 200 - Deep Dive and 300 - Hands On Training. Courses are divided into 1.5 hour blocks, and are offered on Mondays of each SIW and may also be offered during the evening on other days. 

Those interested in obtaining Continuing Education credit for their participation may order a certificate when registering for a workshop.  One certificate for 2.0 Continuing Education Units (CEUs) will be granted per SIW.
2017 Schedule and updates

2017 Fall SIW

Florida Mall Conference Center
Orlando, Florida | 10 - 15 Sep 2017
  0800-1000      SISO 101     HLA 101    VV&A 101      Gateway    Interoperability 101   
  1030-1200     STDS 101       DSEEP 101     DIS 101    Augmented Reality Dev       HLA 202
  1900-2100     FEAT 101    M&S Catalog       DIS 201    Game Engines 101    Agile M&S

Note: A one-time fee of $75.00 allows you to attend any/all tutorials that are offered at this Workshop.

Course catalog: Overview courses

Agile M&S: Maximizing the Efficiency of your M&S Projects

Agile isn't just for software development, it's also well suited to support your needs as it relates to M&S.  Discover how Agile can be used to support the three different bands of M&S: (1) Simulation Development, (2) Simulation Environment Development and Execution, and (3) Standards Development.  The principles of Agile will help you move forward and be even more focused and relevant in supporting the needs of your users and in fostering innovation.

Prerequisite:  A general understanding of modeling and simulation.

Augmented Reality Development

The tutorial provides an overview of development of commercially-available virtual, augmented, and mixed reality devices. Discussion will center on use case demonstrations and relevant integration considerations based on use cases developed at the Air Force Research Laboratory at Wright Patterson Air Force Base in Dayton, Ohio.

The tutorial will cover:
  • Overview of Past, Current, and Future AR and VR devices
    • Market availability and accessibility of commercial products
    • Integration of devices to specific applications
  • Basics of developing with the Microsoft HoloLens using Unity 5.5 and visual studio.
    • Tools needed for development and best practices
    • Short tutorial on basic hologram creation
  • Localization and its requirements including inside out tracking, outside in tracking, and depth sensing
  • Spatial registration and its requirements
  • Project examples include
    • AR NDE app
    • Inhance Digital's F-15 maintainer app
    • (Possible taqtile manifest showcase)

Prerequisite: A general understanding of modeling and simulation.

DIS 101 - Distributed Interactive Simulation

This tutorial provides an overview of:
  • IEEE Std 1278.1TM - 2012, IEEE Standard for Distributed Interactive Simulation - Application Protocols.

The tutorial includes introductory background material describing what Distributed Interactive Simulation (DIS) is; explains why DIS is a viable standard for distributed simulation; and provides information for joining an active developers group. New features added in the 2012 version, including Directed Energy Weapons, Information Operations, and the general extensibility of Protocol Data Units (PDUs) will be discussed.

Prerequisite: A general understanding of modeling and simulation. Minimum technical background needed. Familiarity with distributed real-time simulation of vehicles and weapon system platforms is helpful.

DSEEP 101 - Distributed Simulation Engineering and Execution Process

This tutorial provides an overview of:
  • IEEE Std 1730TM - 2010, IEEE Recommended Practice for Distributed Simulation Engineering and Execution Process (DSEEP)

DSEEP defines the processes and procedures that should be followed by users of distributed simulations to develop and execute their simulations.

The DSEEP generalizes the Federation Development and Execution Process (FEDEP, IEEE 1516.3) to all distributed simulation environments and architectures, no longer focusing solely on the High Level Architecture (HLA).

This tutorial provides the top level steps and supporting activities for the entire process. It also introduces and illustrates the inputs, recommended tasks, and outcomes of the activities.

There will be a brief overview of the architecture-specific annexes for HLA, Distributed Interactive Simulation (DIS), and the Test and Training Enabling Architecture (TENA).

Attendees also will be introduced to:
  • IEEE Std 1730.1TM - 2013, IEEE Recommended Practice for Distributed Simulation Engineering and Execution Process Multi-Architecture Overlay (DMAO)
  • SISO-STD-012-2013, Standard for Federation Engineering Agreements Template

DMAO is the IEEE standard that extends the DSEEP to multi-architecture environments.

Prerequisite: A general understanding of modeling and simulation.

Enterprise Solutions for M&S Asset Discovery and Reuse

Reuse of M&S assets including models, simulations, data, architectures, designs, and requirements (collectively referred to as M&S assets hereafter) is critical to providing cost-effective M&S solutions to meet our mission needs.   The first step in reuse is gain visibility into what assets exist, and then to have mechanisms to access them.  However, most M&S practitioners and managers are largely unaware of the wealth of M&S assets that exist, and they don't know where to go to find them.

Significant advances have been made recently to more effectively support discovery and reuse of assets across the M&S enterprise. This includes both the US DoD and our international partners. Following sound industry practices and DoD guidance, we have migrated to a set of federated discovery and access mechanisms. Those attending this tutorial will learn what types of M&S assets can be discovered, how and where to search for them, and how to obtain access to what they find.   Attendees will also learn how they can make their own assets available to others for reuse.  The tutorial will begin with the motivation and business case for M&S asset reuse, and then through descriptions and demonstrations, show how M&S assets can be discovered and accessed.

Prerequisite:  None.

FEAT 101 – Federation Engineering Agreements Template

The FEAT [SISO-STD-012-2013] is an eXtensible Markup Language (XML) schema that provides a standardized format for recording federation agreements to increase their usability and reuse. The FEAT is designed to:

  • Ensure federation and federate engineers accurately record all necessary agreements by providing a complete list of all potential agreements
  • Ensure agreements are recorded in a standardized, unambiguous manner
  • Enable federation participants to locate agreements rapidly or determine that a required agreement has not been recorded
  • Enforce internal consistency between related agreements
  • Enable the development of a new generation of federation engineering tools driven by the XML schema that reduce the cost, schedule, and risk of developing federations.

This tutorial leverages the FEAT User’s Guide and Fuel Economy Federation FEAT example to illustrate the use of FEAT for a simple federation.

The tutorial will cover:
  • FEAT structure including categories of agreements
  • Design patterns and reuse of existing schemas
  • Supporting tools and documents
  • Examples of applying FEAT to the Fuel Economy Federation

Prerequisite: A general understanding of distributed simulation. An understanding of the Distributed Simulation Engineering and Execution Process (DSEEP) is beneficial.

Game Engines 101

The tutorial includes introductory background material describing the use of video game engines, such as Unity and Unreal, in the development of simulation capabilities for training use cases. Basic game engine mechanics will be covered, and the two foremost game engines will be compared. The tutorial will highlight data access and sharing options, including the data handling native to the engines as well as the use of the Distributed Interactive Simulation (DIS), which enable performance and physiological data and assessment within the simulations. Strategies allowing the integration of external devices and data streams will be explored. Examples will be used to highlight game engine capabilities, data handling, and integration strategies.

The tutorial will cover:
  • Game engine basics including: assets creation & manipulation, physics, scripting, visuals, performance, and plugins
  • Game engine evaluation
  • Data extraction (control input, world information, DIS)
  • Device integration (Strategies, data interleaving)
  • Project examples include:
    • Eye tracker Unity 3D integration and data collection
    • Data collection using Unreal Engine
    • Data stream integration with Unity 3D
    • Virtual environment built in Unreal Engine
    • Multiplayer game-based training simulation using Unity 3D
    • DIS game integration

Prerequisite: A general understanding of modeling and simulation.


Course description is not currently available.

Prerequisite: Not currently available.

Gateways 101 - An Introduction to Gateways in Multi-Architecture Environments

In the distributed simulation world, gateways remain a significant interoperability enabler, particularly in multi-architecture applications which are often used to build Live/Virtual/Constructive (LVC) environments.

This tutorial will serve as an introduction to gateways. Primary learning objectives are:

  • To understand the need/role of gateways in distributed simulation (to better enable interoperability).
  • To understand how gateways operate in a distributed simulation environment.
  • To understand the types of distributed simulation architectures typically involved in Live-Virtual-Constructive (LVC) environments.
  • To understand some of the issues surrounding gateways.
  • To understand how gateways are acquired today.
  • To understand the requirements for configuring and using gateways.
  • To be aware of ongoing work to address some of the issues surrounding gateways.

Prerequisite: A general understanding of modeling and simulation.

GM-VV 101 – An Introduction to the Generic Methodology for Verification and Validation (GM-VV)

The Generic Methodology for Verification and Validation (GM-VV) is a new emerging SISO standard for V&V of M&S assets, and at the same time under consideration by NATO and individual national defence directorates to be incorporated as part of their M&S policies. The GM-VV can be tailored towards the V&V needs of any specific M&S application, technology and organization. The GM-VV provides a comprehensive goal-oriented reasoning network approach to efficiently develop evidence based arguments to justify why M&S assets are acceptable for a specific intended use, or not. This network supports M&S stakeholders in their risk-based decision-making process on the M&S asset development, employment and reuse.

The tutorial will provide the attendees with initial hands-on to support the implementation and execution of V&V within their M&S organization or projects using GM-VV. The tutorial is organized in the following four parts:

  • Part 1 provides the attendees a general introduction to the V&V within the M&S domain. It shows the relationship with V&V practices of common systems and software engineering, and who, where and how one can benefit from using GM-VV.
  • Part 2 introduces the attendees to the three technical frameworks of the GM-VV: 
    • Conceptual framework: provides a set of generic principles and concepts for V&V of M&S assets. 
    • Implementation framework: provides a set of generic V&V product, process and organizational building blocks 
    •  Tailoring framework: provides a set of approaches to develop tailored V&V solutions for M&S assets using the implementation framework building blocks. 
  • Part 3 presents a complete illustration of the GM-VV by means of a real-life case-study example from the training domain. 
  • Part 4 provides the attendees with some basic application guidance for GM-VV in the form of recommended practices, do’s and don’ts, tools and useful referential sources.

HLA 101 – Introduction to High Level Architecture

The High-Level Architecture (HLA) is an international standard for simulation interoperability. It originated in the defense community, but is increasingly used in other domains. This tutorial provides an introduction to HLA and these IEEE Standards that specify HLA:

  • IEEE Std 1516TM - 2010, IEEE Standard for Modeling and Simulation (M&S) High Level Architecture (HLA) – Framework and Rules
  • IEEE Std 1516.1TM - 2010, IEEE Standard for Modeling and Simulation (M&S) High Level Architecture (HLA) – Federate Interface Specification
  • IEEE Std 1516.3TM - 2010, IEEE Standard for Modeling and Simulation (M&S) High Level Architecture (HLA) –Object Model Template (OMT) Specification

This tutorial describes the requirements for interoperability, flexibility, composability and reuse, and how HLA meets those requirements. The principles and terminology of an HLA federation is given including some real world examples.

The tutorial will cover:
  • The HLA Rules that federates and federations follow.
  • The HLA Interface Specification that describes the services a simulation can use for data exchange, synchronization, and overall management.
  • The HLA Object Model Template that is used for describing the data exchange between simulations.

Some practical information is given about current implementations, including commercial-off-the-shelf, government-off-the-shelf, and open source implementations. The continuous development of performance, robustness of the implementations as well as available tools is also described. Finally, some advice is given on how to get started with HLA, including the use of the related process standard:

  • IEEE Std 1730TM - 2010, IEEE Recommended Practice for Distributed Simulation Engineering and Execution Process (DSEEP)

Prerequisite: A general understanding of modeling and simulation.

Interoperability 101

Course description is not currently available.

Prerequisite: A general understanding of modeling and simulation.

Modeling and Simulation of Complex Systems

This tutorial focuses on how to apply of modeling and simulation to engineer and manage large complex systems. It focuses on how M&S can be used to avoid major systemic failures and reduce costs.

The tutorial begins with an overview and examples of complex systems, followed by a survey of several classes of modeling and simulation techniques based on nonlinear differential equations, networks, stochastic models, cellular automata, and swarm-like systems.

Some historical classic large system failures are reviewed, along with demonstrations of how M&S could have prevented them. Existing software models will be used to illustrate systems and provide practical experience.

Challenges for accurately and cost effectively modeling large systems are reviewed, along with recommendations for addressing them while avoiding pitfalls. A survey of M&S frameworks is reviewed. The need for model re-use, and to interface collections of models and simulators is discussed, as well as the use SISO standards to accomplish this efficiently.

Prerequisite:  None.

MURM 101 - M&S Use Risk Methodology

Use of model(s), simulation(s) and the associated data (hereafter referred to as M&S) in development of scientific and technical knowledge; analysis of problems; design, development, and assessment of systems; and support of system operations continues to increase, as does the role that M&S plays in these activities. Hence, it becomes increasingly important to know how much confidence should be placed in M&S results, and what their limits of credibility may be. At present, M&S may be developed and used without a comprehensive appreciation for the uncertainties associated with the M&S and the M&S results. This tutorial presents a mathematically coherent methodology for the assessment of M&S Use Risk that provides an explicit relationship between V&V activities and the risk associated with using M&S results. The methodology is flexible in that it can be employed throughout the M&S lifecycle. By employing Claude E. Shannon's maximum information entropy concept, the methodology helps to preclude unintended bias and allows full use of all available information.

Prerequisite: A general understanding of modeling and simulation and an interest in verification and validation.

OPEN UTF 101 - Introduction to the Open Unified Technical Framework

The Open Unified Technical Framework (OpenUTF) is comprised of three synergistic architectures that are designed to support parallel and distributed computing within a standards-based interoperability framework. The OpenUTF provides a common infrastructure for hosting plug-and-play model and/or service components that are distributed across processors and are able to mutually interact in abstract time (I.e., scaled time or logical time). Because the same framework can be used for operational services and models, the OpenUTF has the potential to unify M&S, service-oriented applications, and T&E.

These architectures are:

  1. Open Modeling and Simulation Architecture (OpenMSA) is a layered architecture, where each layer represents a critical technology for supporting interoperability standards and modern computing on networks of multicore computers. 
  2. Open System Architecture for Modeling and Simulation (OSAMS) is a subset of the OpenMSA. It focuses on modeling constructs and plug-and-play software composability. The goal of OSAMS is to provide standard services that minimize software development efforts while promoting interoperability and reuse of model/service components. Models developed according to the eventual OSAMS standard would be interoperable within any OSAMS-compliant simulation engine.
  3. Open Cognitive Architecture Framework (OpenCAF) extends OSAMS with special modeling constructs for representing intelligent behavior. This includes a reasoning engine that is able to support rule-based, emotion-based, and training-based thought processes that are triggered by external stimulus, along with goal-oriented task management. OpenCAF is necessary to model behaviors of intelligent entities.

This tutorial will provide an introduction to the architectures of the OpenUTF. 

Prerequisite: A general understanding of modeling and simulation.

PDES 101 - Technology of Parallel Discrete Event Simulation

The multicore computing revolution has begun and will change how software is designed, developed, tested, validated, fielded, and maintained. Supporting M&S in parallel and distributed multicore computing environments offers extreme challenges, especially when modeled entities freely interact with each other at any time and/or time scale. This tutorial introduces the core techniques for parallel discrete event simulation (PDES) that have been researched and developed over the past 25 years. These techniques include:

  1. Lock-time stepping
  2. Fixed time windows
  3. Topology-based synchronization
  4. Rollback-based optimistic approaches
  5. The event horizon and risk free optimistic approaches
  6. Hybrid approaches that introduce flow control
  7. Robust repeatability using abstract time representations
  8. Five dimensional simulation

By the end of this tutorial, participants will understand the challenges and general techniques that are used to support PDES. This tutorial will especially help members of SISO understand the core PDES technologies of the OpenUTF.

Prerequisite: A general understanding of modeling and simulation.

Rediscover the Defense M&S Catalog

M&S industry leaders are capitalizing on the new and improved Defense M&S Catalog. This tutorial presents the new approach to search, discover, and publish M&S assets.  It focuses on the purpose of the Defense M&S Catalog and how its new concept of operations simultaneously serves the local needs of M&S practitioners and the policy objectives of the United States Department of Defense.

Live demonstrations will illustrate how to effectively access and utilize the catalog’s upgraded search methods, operate its menu-driven tool for creating valid metadata records, and publish locally-developed metadata records directly to the catalog without data transformation or human intervention.

The tutorial is intended for M&S Leaders, especially those in Systems Engineering, Acquisition, and Program Management, who rely on M&S to develop new system capabilities; as well as developers, project managers, and other senior technical leaders, who are responsible for the management and sharing of M&S assets.

 Prerequisite:  None. 

SEDRIS 101 - An Introduction to SEDRIS Fundamentals

Environmental data is an increasingly integral part of many of today's information technology applications. The methods and techniques for generating, representing, and sharing environmental data play a key role in the interoperation of heterogeneous systems that use such data. SEDRIS is a suite of technologies, standards, implementations, and tools that provide an integrated approach to the representation and interchange of environmental data.

This tutorial highlights the role and importance of standards in the representation, interchange, and reuse of environmental data, and gives an overview of the fundamental concepts and components of SEDRIS. The presentation will touch on how the SEDRIS technology components are used in various applications and in the interchange of environmental data, and will provide an overview of the SEDRIS ISO/IEC standards and their corresponding on-line registries. A brief overview of several key SEDRIS-based tools and utilities will also be included.

Prerequisites: Familiarity with environmental data and concepts, and a fundamental understanding of how models and simulations use and process environmental data.

SISO 101 - Introduction to the SISO Standards Development Process

This tutorial session is hosted by leaders of the SISO Standards Activity Committee who explain SISO’s processes and organizational makeup and role in the development of standards.

Prerequisite: Interest in learning more about SISO and how to become involved in SISO activities.

SOA/LVC 101 - Employing Service Oriented Architecture for Live-Virtual Constructive Multi-Architecture Distributed Simulations

Building Live-Virtual-Constructive (LVC) multi-architecture distributed simulations upon Service Oriented-Architecture (SOA) has been demonstrated and studied, but has not been embraced in general by the community served by modeling and simulation (M&S), or LVC multi-architecture developers. This is due to the real and perceived up-front costs of employing a new technology to address compatibility issues that have traditionally been addressed with ad hoc gateways and bridges, one-of-a-kind database connectors, and other single-point design solutions.

While SOA will not directly address composability of multiple simulations, nor eliminate the need for gateways and bridges, it can be a critical component in integration and management. It also has the potential to provide rapid deployment of integrated M&S components.

The intent of this tutorial is to present a balanced view of the considerations for using SOA as an M&S architecture. The tutorial provides an overview of SOA concepts, the challenges of integrating LVC multi-architecture distributed simulations, an explanation of the benefits and barriers to developing/integrating multi-architecture distributed simulations into a SOA construct, and when and when not to attempt to use SOA as a long-range infrastructure for M&S integration.

  • Introduction to SOA Concepts 
  • Overview of LVC Multi-Architecture Distributed Simulations 
  • Execution of LVC Distributed Simulation in a SOA Construct Architecture 
  • Design Perspectives of SOA for M&S 
  • The Issues & Challenges, Benefits & Barriers 
  • Overview of the Current State of DoD SOA Services 
  • Recent Examples of Successes and Problems in using SOA for M&S 
  • When and When Not to Attempt to Use a SOA-Based M&S Architecture

Prerequisite: A general understanding of modeling and simulation.

SPACE Smackdown 101 

At the Spring 2011 Simulation Interoperability Workshop (SIW), the Space Forum sponsored a "Space Smackdown" event which sought to increase awareness of HLA in the academic simulation community.  It was an outreach of sorts, an effort to expand the HLA community beyond its current base.

The event involved teams from several universities. Each team built one or more space vehicle federates which joined a simulated Earth-Moon federation governed by a single HLA-Evolved object model (FOM). Although the scope of the simulation was modest, its objectives of introducing a new community to HLA were quite successfully met.

This tutorial will discuss the Space Smackdown with a particular focus on the upcoming smackdown event at the Spring 2012 SIW.  We will cover the following topics:

  • An introduction to the motivations behind the Space Smackdown,
  • A review of the scenario that was simulated in the 2011 smackdown, including the HLA-Evolved FOM modules that the federates adhered to,
  • A summary of the highlights of the 2011 event, including the technology challenges, overall results and lessons learned, and
  • A discussion of current plans for the 2012 event, including a brainstorming discussion of possible mission scenarios.

The tutorial is intended for people curious about or interested in participating in the upcoming 2012 smackdown. We will cover the prerequisites for teams wishing to participate, and we will let people know how to get involved.

Prerequisite: A general understanding of modeling and simulation.

STDS 101 - SISO Standards: An Introduction to the SISO Standards Development Process

This tutorial session is hosted by leaders of the SISO Standards Activity Committee who explain how to become involved in the SISO standards development and support process. Note: No tutorial fee is charged for this session.

Prerequisite: A general understanding of modeling and simulation.


The Test and Training Enabling Architecture (TENA) provides an advanced set of interoperability software and interfaces for use in joint distributed testing and training. The TENA software includes the TENA Middleware, a high-performance, real-time, low-latency communication infrastructure used by training range instrumentation software and tools during execution of a range training event. The standard TENA Object Model provides data definitions for common range entities and thus enables semantic interoperability among training range applications. The TENA tools, utilities, and gateways assist in creating and managing an integration of range resources. The current version of the TENA Middleware, Release 6.0.2, is being used by the range community for testing, training, evaluation, and feedback and will be used in major exercises in the future.

The Joint Mission Environment Test Capability (JMETC) program is chartered to create a persistent test and evaluation capability throughout the US DOD. JMETC consists of a persistent network; a set of TENA-compliant software middleware, interfaces, tools, and databases; and a process for creating large distributed test events. The combination of TENA and JMETC gives testers and trainers unprecedented power to craft a joint distributed mission environment that meets testing and training requirements for the warfighter.

Prerequisite: A general understanding of modeling and simulation and an interest in testing and training.

The Certified M&S Professional Program (www.simprofessional.org)

The recognition of M&S as a profession and its practitioners as professionals requires several things. Key among them is the Certified Modeling and Simulation Professional Development Program (CMSP), which is administered by the Modeling and Simulation Professional Certificate Council. This program, co-sponsored by SISO, SCS, and NTSA, has been in existence since 2008, but has just undergone a refresh to ensure test content is current, the processes are clear, and the program is credible. The purpose of this course is to describe the program, its role in the M&S profession, its benefits to certificate holders and to the community, and to provide a test preparation class. The class provides a top level overview of the broad range of subject matter covered by the examination, and includes discussion of sample questions drawn from the actual question bank. The ongoing validation and improvement process will be emphasized.

At the completion of the class the student will be able to describe the CMSP program including the requirements, application process, testing process, certificate renewal process, and the two types of certificate (Technical / Developer and Manager / User). A certificate for 1 CEU (Continuing Education Unit) will be awarded.

Prerequisite: The desire and ability to be recognized as a professional M&S practitioner.

VV&A 101 - Verification, Validation, & Accreditation

This tutorial provides an overview of the processes of Verification, Validation, and Accreditation, which are foundational elements that underlie assessments of M&S credibility. Information derived from the VV&A processes is used to shape the understanding of how and where an M&S should be used and under what the constraints.

While VV&A is founded on basic software engineering principles, implementation is often constrained by resources, whether these resources be time, money, personnel, or information. This tutorial will introduce M&S Users, M&S Developers, and VV&A Practitioners to the key concepts associated with VV&A planning and implementation, the impacts and the drivers, and basic documentation requirements.

Prerequisite: A general understanding of modeling and simulation.

Course catalog: Deep dive courses

DIS 201 - New Extensibility and Dead Reckoning Features in DIS Version 7

This tutorial provides an in-depth review of two features: PDU extensibility and improvements in dead reckoning.

PDU extensibility expands the ability of DIS users to add custom data to PDUs. Some PDUs allow user-defined records to be directly added. Other PDUs can be extended using the new Attribute PDU. Both methods retain compatibility with older versions of DIS. This allows customized PDUs to be added in new or upgraded simulations while maintaining interoperability to older simulations that cannot be modified.

Dead reckoning has been enhanced in DIS Version 7, mainly in the extrapolation of entity orientation. A new geometric method of determining the orientation threshold is described using either quaternions or rotation matrices. This method avoids the problems of Euler angle singularities than can cause excessively high PDU transmit rates. Other new features speed up dead reckoning calculations in receiving simulations by adding extra information in the Entity State PDU. These new features maintain full backward and forward compatibility with DIS Version 5 and 6.

Prerequisite: Familiarity with basic DIS PDU usage. Knowledge of dead reckoning is helpful but the tutorial will include an introduction to the concepts.

Gateways 201 - Technical Issues with Gateway Selection and Configuration

The Gateways 201 tutorial is designed to build upon the introductory Gateways Tutorial by diving deeper into the technical issues associated with gateway selection and configuration along with potential solutions to those issues.  The tutorial will begin with a brief review of the types of problems encountered by gateway users today and then examine ongoing work within the LVC Architecture Roadmap Implementation (LVCAR-I) and SISO related to gateways.

 Primary learning objectives are:

  • To understand the role of gateways in distributed simulation environments
  • To better understand the problems related to employing gateways
  • To better understand potential solutions to these problems
  • To understand current work within SISO (GDACL PDG) to define a set of standardized gateway languages to assist gateway users:  

                 - Gateway Description Language
                 - SML – SDEM Mapping Language
                 - GFL – Gateway Filtering Language

  • Understand how gateway performance metrics are being addressed.

 Prerequisite: A general understanding of modeling and simulation.

HLA 201 - HLA Evolved — An Overview

This tutorial gives an overview of the new features of HLA Evolved (IEEE 1516-2010) which is a superset of the previous HLA 1516-2000 standard. It describes the new functionality and what new capabilities it provides to federations. It also gives an overview of the open standardization process behind this new version. Some key new features include Modular FOMs, extended XML features, Fault Tolerance, Dynamic Link Compatibility, Encoding helpers, Web Services and Smart Update Rate Reduction.

Finally some approaches for migrating existing federations to HLA 1516-2010 are given, including notes on tool support. An extensive list of in-depth reading is also provided.

Note that a short overview of FOM Modules is included but participants may choose to attend HLA 202 for a detailed walkthrough.

Prerequisites: General knowledge about the purpose and principles of HLA and HLA 101.

HLA 202 - Developing Federation Object Models with HLA Evolved and Beyond

This tutorial provides an overview of how to apply:
  • IEEE Std 1516.3TM - 2010, IEEE Standard for Modeling and Simulation (M&S) High Level Architecture (HLA) –Object Model Template (OMT) Specification
  • Additional features for FOM development, provided in the upcoming version of HLA, informally called “HLA 4”.

The Federation Object Model (FOM) specifies the data that is exchanged between federates in an HLA federation. This tutorial shows the principles of good FOM design. It describes how FOMs can be developed maintained and reused using the Modular FOM concept, introduced in HLA Evolved (IEEE 1516-2010).

Special attention is given to the modular version of the SISO Real-time Platform Reference Federation Object Model (RPR FOM) 2.0. Other FOMs covered are the NATO Education and Training Network FOM and the upcoming SISO Space Reference FOM.

Finally, it presents new features for developing FOMs that are proposed for the next version of HLA, in particular for extending reference FOMs.

Prerequisites: General knowledge about the purpose and principles of HLA corresponding to HLA 101. Some experience with HLA object model development is useful, but not required.

SEDRIS 201 - Using SEDRIS Software and Tools

A fundamental objective in SEDRIS is the representation of complex environmental data and the seamless interchange of environmental data sets. This tutorial is geared toward software developers and environmental data modelers seeking an overview of the SEDRIS software development kits (SDK) and associated tools for accessing, inspecting, and manipulating environmental data.

The tutorial will show how the SEDRIS SDK is used to create applications and libraries that can read and write SEDRIS transmittals. The presentation will demonstrate the process of obtaining and setting up the appropriate SEDRIS component SDKs, depending on the needs of the application.

The tutorial will also show how the SEDRIS tools are used to convert and integrate databases to/from such data formats as Shapefile, GeoTIFF, CTDB, and others. The presentation will cover aspects of verifying the SEDRIS transmittals for conformance to the syntax and rules of the SEDRIS data representation model (DRM), and steps for creating and integrating databases using the Focus tool.

Prerequisites: General knowledge of SEDRIS concepts and components, familiarity with software development and use in environmental data generation and consumption.

VV&A 201: Key Drivers to Efficient VV&A Implementation

The objective of this tutorial is to provide those interested in the planning and implementation of VV&A with guidance on how to address key implementation issues and challenges. Topics to be covered include:

  • How requirements traceability enhances the VV&A processes
  • How to derive "acceptable" acceptability criteria
  • How risk-based tailoring can impact VV&A planning and implementation
  • How to manage and document the V&V test process
  • How to use MIL-STD 3022 (Documentation of VV&A for M&S)

Prerequisites: General knowledge about the purpose and principles of VV&A corresponding to VV&A 101.


OpenUTF - Hands-on Training for the Open Technical Framework

This hands-on course includes seven 1.5-hour sessions over three days. Participants are strongly encouraged to participate in all sessions. See requirements below regarding license agreements and participation of non-US citizens.

The OpenUTF is an emerging framework for hosting next-generation, composable, scalable, parallel and distributed, M&S systems. It is comprised of three synergistic architectures.

  1. Open Modeling and Simulation Architecture (OpenMSA) is a layered architecture, where each layer represents a critical technology for supporting interoperability standards and modern parallel and distributed computing on networks of multicore computers.
  2. Open System Architecture for Modeling and Simulation (OSAMS) is a subset of the OpenMSA. It focuses on modeling constructs that are designed to support plug-and-play software composability. OSAMS provides a programming framework that minimizes software development efforts while promoting interoperability and reuse of plug-and-play model/service components.
  3. Open Cognitive Architecture Framework (OpenCAF) extends OSAMS with modeling constructs for representing intelligent behavior. This includes a reasoning engine that is able to support rule-based, emotion-based, and training-based thought processes, along with goal-oriented task management.

The WarpIV Kernel provides the open-source reference implementation of the OpenUTF core infrastructure and is made freely available to all qualifying United States and Canadian organizations for non-commercial use. Non-U.S. citizens are permitted to participate in this training event, but would require an export license to receive a copy of the software. This training event will utilize the WarpIV Kernel in its hands-on assignments; each designed to guide participants through the primary modeling constructs of the OpenUTF. All training materials, including a quick reference guide and a set of assignment worksheets, will be provided to participants. By the end of this training event, participants will become familiar with the OpenUTF modeling constructs and be able to develop parallel and distributed simulations on their own.

Prerequisites: No parallel or distributed computing experience is required for this training event. However, participants should be somewhat familiar with C++ and/or basic programming concepts. The class will be broken up into small teams of 3-4 participants for the hands-on assignments, with at least one strong lead programmer per team. The instructor will guide the teams through each hands-on exercise. The instructor will provide laptops for use by participants as available. U.S. and Canadian participants wishing to use their own laptops must apply for and receive a software license prior to the training.