An overview of the reporting capabilities in Team Foundation Server 2012. Includes explanation of Data warehouse architecture, included reports (SSRS and Excel), and customization.
Team Foundation Server - Tracking & ReportingSteve Lange
Comprehensive presentation detailing reporting and tracking capabilities of Team Foundation Server. Focuses on Excel workbooks and Reporting Services, but touches on other technologies as well.
Visual Studio ALM 2013 - Edition ComparisonSteve Lange
A comparison of the various editions of Visual Studio 2013, specifically focused on ALM capabilities. It excludes discussion specific to Visual Studio Online subscriptions.
A full overview of Team Foundation Server 2010 (not just what's new).
Includes 4 main areas:
- Manage & Plan your Project
- Understand Parallel Development
- No More "No Repro" Bugs
- Reporting on your Entire Portfolio.
Screenshots are included.
Team Foundation Server Process Templates For Effective Project ManagementAaron Bjork
An introduction to process templates, an overview of the Microsoft Process Template, and new process template features shipping in Team Foundation Server 2010.
Implementing Scrum with Microsoft Team Foundation Service (TFS)Aspenware
Day one Implementing Scrum with Microsoft Team Foundation Service (TFS) training covering the following topics:
TFS Overview
TFS Version Comparison and Installation
Setting Up Your Code in TFS Source Control
Setting Up Your Code in Git Source Control
Scrum Overview
Sprint 0 Activities
Sprint Planning Exercise
Summary and Wrap Up
Team Foundation Server - Tracking & ReportingSteve Lange
Comprehensive presentation detailing reporting and tracking capabilities of Team Foundation Server. Focuses on Excel workbooks and Reporting Services, but touches on other technologies as well.
Visual Studio ALM 2013 - Edition ComparisonSteve Lange
A comparison of the various editions of Visual Studio 2013, specifically focused on ALM capabilities. It excludes discussion specific to Visual Studio Online subscriptions.
A full overview of Team Foundation Server 2010 (not just what's new).
Includes 4 main areas:
- Manage & Plan your Project
- Understand Parallel Development
- No More "No Repro" Bugs
- Reporting on your Entire Portfolio.
Screenshots are included.
Team Foundation Server Process Templates For Effective Project ManagementAaron Bjork
An introduction to process templates, an overview of the Microsoft Process Template, and new process template features shipping in Team Foundation Server 2010.
Implementing Scrum with Microsoft Team Foundation Service (TFS)Aspenware
Day one Implementing Scrum with Microsoft Team Foundation Service (TFS) training covering the following topics:
TFS Overview
TFS Version Comparison and Installation
Setting Up Your Code in TFS Source Control
Setting Up Your Code in Git Source Control
Scrum Overview
Sprint 0 Activities
Sprint Planning Exercise
Summary and Wrap Up
Introduction to Team Foundation Server (TFS) OnlineDenis Voituron
TFS est la plateforme de collaboration qui se trouve au coeur de la solution de gestion du cycle de vie des applications (ALM) de Microsoft. Pour de petites équipes de développement (5 users), TFS est disponible online et gratuitement.
Lors de cette session, principalement à base d'exemples pratiques, nous aborderons les modules Source Controle, Collaborate (gestion des tâches et des bugs) et Automatic Builds (compilations et déploiements automatisés).
This presentation gives you a detailed look at what is in the out of the box templates available in TFS 2013, how they differ, and how that affects some of the ALM tooling.
Implementing Scrum with Microsoft Team Foundation Service (TFS)Aspenware
This training will provide a deep dive into Microsoft Team Foundation Server (TFS) for agile projects from setting up TFS through the end of the first sprint. This is a hands on training, attendees will actively engaged in a sample project using TFS in the cloud. Presenters will include senior Aspenware architects and project managers as well as Steve Lange, Developer Technology Specialist at Microsoft. This training is appropriate for developers, project managers and business analysts. A basic understanding of scrum and agile development is required.
Part 2 Agenda
*Brief review of Part I
*Reporting and tracking features of TFS.
*Setup Continuous Integration and discuss value
*Setup an auto deployment to Azure
*Testing features of TFS and how auto deployments aid that process
*End of Sprint Demo
*End of Sprint Retrospective
*Use TFS to review the tasks and determine velocity on this Sprint
*How to plan subsequent sprints
Learn how to use Devops from beginner level to advanced techniques which is taught by experienced working professionals. With our Devops Training in Chennai you’ll learn concepts in expert level with practical manner.
SCRUM has grown in popularity and acceptance by many companies over the world with numbers of registered SCRUM Masters reaching 51,955 (11 March 2009 - Jeff Sutherland). Although SCRUM does not stipulate what tools to use to produce the necessary artifacts, Microsoft Team Foundation System provides a number of features via TFS Explorer that facilitate capturing the artifacts of SCRUM and is a useful tool for any SCRUM Master, Team and Product Owner. This presentation will highlight the SCRUM framework and show you practical use of TFS and other tools that facilitate the ceremonies and artifacts of SCRUM.
Agile project management with visual studio tfs 2013 - My presentation at Reg...Om Prakash Bang
This presentation is to give overview of Agile Planning for continuous delivery for value, Agile Project Management dashboard, Sprint Planning, Burn Down Charts. Distributed project team collaborate using Team Room, Task board update for all work and used during Daily Stand Up meeting.
Out of box (OOB) template for CMMI, Agile and Scrum. The main distinctions between the three default process templates are in the work item types they provide for planning and tracking work. Visual Studio Scrum is the most light-weight and MSF for Capability Maturity Model Integration (CMMI) provides the most support for formal processes and change management.
1. Microsoft Visual Studio Scrum 2013 - Choose Visual Studio Scrum if your team manages bugs along with product backlog items during sprint planning.
2. MSF for Agile Software Development 2013 - Choose Agile if your organization triages bugs separately from the product backlog and resolves work items before closing them. Also, choose Agile if your team allocates time for bugs with each sprint.
3. MSF for CMMI Process Improvement 2013 - Choose CMMI if your organization triages bugs separately from the product backlog, resolves work items before closing them, and tracks changes to requirements formally. The CMMI template is designed to support formal change management processes.
Visual Studio 2015 is going to be a huge change for both windows and non-windows developers. Thanks to a new/refreshing/cool/awesome change of attitude, Microsoft is embracing the winning tools in the development space and is building them into the next version of Visual Studio.
The open sourcing of .NET Core 5 and ASP.NET 5 along with the release of the free Visual Studio Community Edition and the upcoming release of Windows 10 for all devices makes for some very exciting times ahead.
#ATATalk - Episode 1 : Session on Selenium Exceptions by Pallavi SharmaAgile Testing Alliance
Pallavi Sharma delivered a session on Selenium Exceptions at #ATATalk - Episode 1
Pallavi Sharma is acting as a coach, writer, speaker and owner at 5 Elements Learning where she collaborate and work with many learning enthusiasts and mentors around the globe and use her polyglot skills. An avid learner, she likes to keep herself updated to latest trends and technologies by attending conferences and workshops, and life has been kind enough to provide her with such experiences.
To know more about ATA Events please check : https://ataevents.org/
Relieveing the Testing Bottle Neck - WebinarCprime
When shifting to Agile, testing is often a bottleneck in the process, as it is the last step in the cycle. But, the responsibility to remove the bottleneck is not on the tester alone.
DevOps and Continuous Delivery with Visual Studio 2015 and VSTSSolidify
Som utvecklare är det ofta självklart att kodandet är den viktigaste aktiviteten i projektet. Men om inte funktioner för releaseplanering, källkodshantering, deployment och testning finns på plats så är det vanligt att dessa aktiviteter tar en stor del av utvecklingsarbetet.
DevOps och Continuous Delivery är synsätt som vi använder för att säkerställa snabba leveranser av god kvalitet. Microsoft Azure och Visual Studio har mängder av funktioner som underlättar arbetet så att arbetssätt och verktyg samverkar på ett effektivt sätt.
I detta seminarium kommer vi att förklara vad DevOps är för något och hur detta synsätt kan användas för att säkerställa snabba leveranser med god kvalitet. DevOps hjälper oss överbrygga arbetet i utveckling och drift med fokus på att säkerställa leverenspresision och kvalitet, bland annat genom att se till att bra lösningar för applikationsövervakning, felrapportering, paketering och deployment finns på plats.
Den andra delen vi kommer fokusera på är Continuous Delivery. Med den senaste versionen av Visual Studio Release Management visar vi vad Continuous Delivery är genom att sätta upp en komplett lösning som automatiskt tar kod till produktion. I Visual Studio 2015 sviten har vi ett helt nytt byggsystem och väl integrerade lösningar kring release management, deployment och automatiserad testning som gör plattformen förträfflig för att implementera en continuous delivery process. För att så snabbt och enkelt som möjligt tillgodogöra oss dessa koncept kommer vi använda molnplattformarna Azure och Visual Studio Online, med dessa kan vi flytta utvecklingsplattformen till molnet och få helt nya förutsättningar för snabbare releaser.
Så missa inte denna chans att vara en av de första att se en komplett Continuous Delivery-lösning byggd på den senaste Microsofttekniken!
Introduction to Team Foundation Server (TFS) OnlineDenis Voituron
TFS est la plateforme de collaboration qui se trouve au coeur de la solution de gestion du cycle de vie des applications (ALM) de Microsoft. Pour de petites équipes de développement (5 users), TFS est disponible online et gratuitement.
Lors de cette session, principalement à base d'exemples pratiques, nous aborderons les modules Source Controle, Collaborate (gestion des tâches et des bugs) et Automatic Builds (compilations et déploiements automatisés).
This presentation gives you a detailed look at what is in the out of the box templates available in TFS 2013, how they differ, and how that affects some of the ALM tooling.
Implementing Scrum with Microsoft Team Foundation Service (TFS)Aspenware
This training will provide a deep dive into Microsoft Team Foundation Server (TFS) for agile projects from setting up TFS through the end of the first sprint. This is a hands on training, attendees will actively engaged in a sample project using TFS in the cloud. Presenters will include senior Aspenware architects and project managers as well as Steve Lange, Developer Technology Specialist at Microsoft. This training is appropriate for developers, project managers and business analysts. A basic understanding of scrum and agile development is required.
Part 2 Agenda
*Brief review of Part I
*Reporting and tracking features of TFS.
*Setup Continuous Integration and discuss value
*Setup an auto deployment to Azure
*Testing features of TFS and how auto deployments aid that process
*End of Sprint Demo
*End of Sprint Retrospective
*Use TFS to review the tasks and determine velocity on this Sprint
*How to plan subsequent sprints
Learn how to use Devops from beginner level to advanced techniques which is taught by experienced working professionals. With our Devops Training in Chennai you’ll learn concepts in expert level with practical manner.
SCRUM has grown in popularity and acceptance by many companies over the world with numbers of registered SCRUM Masters reaching 51,955 (11 March 2009 - Jeff Sutherland). Although SCRUM does not stipulate what tools to use to produce the necessary artifacts, Microsoft Team Foundation System provides a number of features via TFS Explorer that facilitate capturing the artifacts of SCRUM and is a useful tool for any SCRUM Master, Team and Product Owner. This presentation will highlight the SCRUM framework and show you practical use of TFS and other tools that facilitate the ceremonies and artifacts of SCRUM.
Agile project management with visual studio tfs 2013 - My presentation at Reg...Om Prakash Bang
This presentation is to give overview of Agile Planning for continuous delivery for value, Agile Project Management dashboard, Sprint Planning, Burn Down Charts. Distributed project team collaborate using Team Room, Task board update for all work and used during Daily Stand Up meeting.
Out of box (OOB) template for CMMI, Agile and Scrum. The main distinctions between the three default process templates are in the work item types they provide for planning and tracking work. Visual Studio Scrum is the most light-weight and MSF for Capability Maturity Model Integration (CMMI) provides the most support for formal processes and change management.
1. Microsoft Visual Studio Scrum 2013 - Choose Visual Studio Scrum if your team manages bugs along with product backlog items during sprint planning.
2. MSF for Agile Software Development 2013 - Choose Agile if your organization triages bugs separately from the product backlog and resolves work items before closing them. Also, choose Agile if your team allocates time for bugs with each sprint.
3. MSF for CMMI Process Improvement 2013 - Choose CMMI if your organization triages bugs separately from the product backlog, resolves work items before closing them, and tracks changes to requirements formally. The CMMI template is designed to support formal change management processes.
Visual Studio 2015 is going to be a huge change for both windows and non-windows developers. Thanks to a new/refreshing/cool/awesome change of attitude, Microsoft is embracing the winning tools in the development space and is building them into the next version of Visual Studio.
The open sourcing of .NET Core 5 and ASP.NET 5 along with the release of the free Visual Studio Community Edition and the upcoming release of Windows 10 for all devices makes for some very exciting times ahead.
#ATATalk - Episode 1 : Session on Selenium Exceptions by Pallavi SharmaAgile Testing Alliance
Pallavi Sharma delivered a session on Selenium Exceptions at #ATATalk - Episode 1
Pallavi Sharma is acting as a coach, writer, speaker and owner at 5 Elements Learning where she collaborate and work with many learning enthusiasts and mentors around the globe and use her polyglot skills. An avid learner, she likes to keep herself updated to latest trends and technologies by attending conferences and workshops, and life has been kind enough to provide her with such experiences.
To know more about ATA Events please check : https://ataevents.org/
Relieveing the Testing Bottle Neck - WebinarCprime
When shifting to Agile, testing is often a bottleneck in the process, as it is the last step in the cycle. But, the responsibility to remove the bottleneck is not on the tester alone.
DevOps and Continuous Delivery with Visual Studio 2015 and VSTSSolidify
Som utvecklare är det ofta självklart att kodandet är den viktigaste aktiviteten i projektet. Men om inte funktioner för releaseplanering, källkodshantering, deployment och testning finns på plats så är det vanligt att dessa aktiviteter tar en stor del av utvecklingsarbetet.
DevOps och Continuous Delivery är synsätt som vi använder för att säkerställa snabba leveranser av god kvalitet. Microsoft Azure och Visual Studio har mängder av funktioner som underlättar arbetet så att arbetssätt och verktyg samverkar på ett effektivt sätt.
I detta seminarium kommer vi att förklara vad DevOps är för något och hur detta synsätt kan användas för att säkerställa snabba leveranser med god kvalitet. DevOps hjälper oss överbrygga arbetet i utveckling och drift med fokus på att säkerställa leverenspresision och kvalitet, bland annat genom att se till att bra lösningar för applikationsövervakning, felrapportering, paketering och deployment finns på plats.
Den andra delen vi kommer fokusera på är Continuous Delivery. Med den senaste versionen av Visual Studio Release Management visar vi vad Continuous Delivery är genom att sätta upp en komplett lösning som automatiskt tar kod till produktion. I Visual Studio 2015 sviten har vi ett helt nytt byggsystem och väl integrerade lösningar kring release management, deployment och automatiserad testning som gör plattformen förträfflig för att implementera en continuous delivery process. För att så snabbt och enkelt som möjligt tillgodogöra oss dessa koncept kommer vi använda molnplattformarna Azure och Visual Studio Online, med dessa kan vi flytta utvecklingsplattformen till molnet och få helt nya förutsättningar för snabbare releaser.
Så missa inte denna chans att vara en av de första att se en komplett Continuous Delivery-lösning byggd på den senaste Microsofttekniken!
Working as a Team with Team Foundation Server 2010gueste6f731
This presentation highlights some of the new features in Team Foundation Server 2010 and how they help improve development environments for small to enterprise teams.
The ultimate presentation about Scrum, the world's leading project management framework for agile software development.
http://www.noop.nl
http://www.jurgenappelo.com
Начиная с 2006 года мы провели
более 400 проектов, связанных
с исследованием пользователей,
юзабилити-тестированием,
информационной архитектурой,
проектированием экранов и другими
способами улучшения интерфейсов
программных продуктов.
Мы занимаем более 50%
рынка в сфере
юзабилити-услуг в России.
Just uploaded this for someone but i see that animations are broken. The slide is also not that usefull currently as it requires some more explanations.
Releasing fast code - The DevOps approachMichael Kopp
Agile makes you Develop faster, DevOps also makes you Deploy faster but how do you make your Application faster?
Many currently used Performance Management practices don’t work anymore as they are too time consuming. It takes a new approach to track performance in Continuous Integration, get more value out of Load Testing and leverage production data for performance optimization.
We will show you real world examples on how the new DevOps approach can work.
Moving 65,000 Microsofties to DevOps with Visual Studio Team ServicesVSTS Community MSFT
How do you migrate over 65,000 of the most demanding software engineers from infrastructure built up over decades of high-intensity work into a common engineering system based on modern software development technologies and best practices?
If you had an opportunity to build an application from the ground up, with testability a key design goal, what would you do?
In this presentation, we will look at just such a situation - a major, two year rewrite of a suite of core business systems. We will discuss how a system looks when testability is as important as functionality - and what it looks like when quality concerns are part of the initial design. We will look at the role of test automation and manual test in a modern project, and look at the tools and processes. The session will conclude with a demo of the latest visual test automation tool from MIT and a Q&A.
Codesion's Live Webinar: Cutting Edge Cloud Development Tools for Force.comCollabNet
Would you like to witness the industry's first end-to-end cloud development solution for Force.com? For the first time see version control, project planning, project management, and one-click deployment integrated into a single environment, designed specifically for teams building for the Force.com platform. Join Brian Matthews, the founder and CTO of BrainEngine, and Willie Wang, VP Product and Services of Codesion Cloud Services, as they introduce a new and innovative way to build and deploy cloud applications for Force.com.
DevOps in the Amazon Warehouse - Shawn GandhiTriNimbus
Presentation from Toronto's 2016 Canadian Executive Cloud & DevOps Summit on Friday, November 4th.
Speaker: Shawn Gandhi, Head of Solutions Architecture, AWS Canada
Title: Rogue Development: DevOps in the Amazon Warehouse
Similar to Team Foundation Server 2012 Reporting (20)
Generating a custom Ruby SDK for your web service or Rails API using Smithyg2nightmarescribd
Have you ever wanted a Ruby client API to communicate with your web service? Smithy is a protocol-agnostic language for defining services and SDKs. Smithy Ruby is an implementation of Smithy that generates a Ruby SDK using a Smithy model. In this talk, we will explore Smithy and Smithy Ruby to learn how to generate custom feature-rich SDKs that can communicate with any web service, such as a Rails JSON API.
Essentials of Automations: Optimizing FME Workflows with ParametersSafe Software
Are you looking to streamline your workflows and boost your projects’ efficiency? Do you find yourself searching for ways to add flexibility and control over your FME workflows? If so, you’re in the right place.
Join us for an insightful dive into the world of FME parameters, a critical element in optimizing workflow efficiency. This webinar marks the beginning of our three-part “Essentials of Automation” series. This first webinar is designed to equip you with the knowledge and skills to utilize parameters effectively: enhancing the flexibility, maintainability, and user control of your FME projects.
Here’s what you’ll gain:
- Essentials of FME Parameters: Understand the pivotal role of parameters, including Reader/Writer, Transformer, User, and FME Flow categories. Discover how they are the key to unlocking automation and optimization within your workflows.
- Practical Applications in FME Form: Delve into key user parameter types including choice, connections, and file URLs. Allow users to control how a workflow runs, making your workflows more reusable. Learn to import values and deliver the best user experience for your workflows while enhancing accuracy.
- Optimization Strategies in FME Flow: Explore the creation and strategic deployment of parameters in FME Flow, including the use of deployment and geometry parameters, to maximize workflow efficiency.
- Pro Tips for Success: Gain insights on parameterizing connections and leveraging new features like Conditional Visibility for clarity and simplicity.
We’ll wrap up with a glimpse into future webinars, followed by a Q&A session to address your specific questions surrounding this topic.
Don’t miss this opportunity to elevate your FME expertise and drive your projects to new heights of efficiency.
Software Delivery At the Speed of AI: Inflectra Invests In AI-Powered QualityInflectra
In this insightful webinar, Inflectra explores how artificial intelligence (AI) is transforming software development and testing. Discover how AI-powered tools are revolutionizing every stage of the software development lifecycle (SDLC), from design and prototyping to testing, deployment, and monitoring.
Learn about:
• The Future of Testing: How AI is shifting testing towards verification, analysis, and higher-level skills, while reducing repetitive tasks.
• Test Automation: How AI-powered test case generation, optimization, and self-healing tests are making testing more efficient and effective.
• Visual Testing: Explore the emerging capabilities of AI in visual testing and how it's set to revolutionize UI verification.
• Inflectra's AI Solutions: See demonstrations of Inflectra's cutting-edge AI tools like the ChatGPT plugin and Azure Open AI platform, designed to streamline your testing process.
Whether you're a developer, tester, or QA professional, this webinar will give you valuable insights into how AI is shaping the future of software delivery.
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf91mobiles
91mobiles recently conducted a Smart TV Buyer Insights Survey in which we asked over 3,000 respondents about the TV they own, aspects they look at on a new TV, and their TV buying preferences.
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
Key Trends Shaping the Future of Infrastructure.pdfCheryl Hung
Keynote at DIGIT West Expo, Glasgow on 29 May 2024.
Cheryl Hung, ochery.com
Sr Director, Infrastructure Ecosystem, Arm.
The key trends across hardware, cloud and open-source; exploring how these areas are likely to mature and develop over the short and long-term, and then considering how organisations can position themselves to adapt and thrive.
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
Accelerate your Kubernetes clusters with Varnish CachingThijs Feryn
A presentation about the usage and availability of Varnish on Kubernetes. This talk explores the capabilities of Varnish caching and shows how to use the Varnish Helm chart to deploy it to Kubernetes.
This presentation was delivered at K8SUG Singapore. See https://feryn.eu/presentations/accelerate-your-kubernetes-clusters-with-varnish-caching-k8sug-singapore-28-2024 for more details.
Connector Corner: Automate dynamic content and events by pushing a buttonDianaGray10
Here is something new! In our next Connector Corner webinar, we will demonstrate how you can use a single workflow to:
Create a campaign using Mailchimp with merge tags/fields
Send an interactive Slack channel message (using buttons)
Have the message received by managers and peers along with a test email for review
But there’s more:
In a second workflow supporting the same use case, you’ll see:
Your campaign sent to target colleagues for approval
If the “Approve” button is clicked, a Jira/Zendesk ticket is created for the marketing design team
But—if the “Reject” button is pushed, colleagues will be alerted via Slack message
Join us to learn more about this new, human-in-the-loop capability, brought to you by Integration Service connectors.
And...
Speakers:
Akshay Agnihotri, Product Manager
Charlie Greenberg, Host
17. Excel Reports Excel Report
Project Management Generation
Bug Backlog Management Create reports in Excel from work
item queries
Build Management
Test Management
18.
19.
20.
21.
22.
23.
24.
25. Quickly find important information about team projects
Show project data, support investigation, & help teams perform common tasks more quickly.
26. Windows SharePoint SharePoint Server SharePoint Server
Services 3.0 Standard Enterprise
My
X X X
Project
X X
Progress
X
Quality
X
Test
X
Bugs
X
Build
X
27. What is the next set of Tasks, Bugs, or Test Cases that I should act on?
What is the status of the team's most recent builds?
28. Is the team likely to finish the iteration on time?
Will the team complete the planned work based on the current
burn rate?
What were the most recent check-ins?
Burn Rate
Work Item Breakdown
Burndown
29. Is the team likely to finish the iteration on time?
Will the team complete the planned work based on the current
burndown?
How much progress has the team made on implementing user
stories in the past four weeks?
How quickly is the team identifying and closing Issues?
What were the most recent check-ins?
30. Is the test effort on track?
Is the team testing the appropriate functionality?
Are the team's bug fixes of high quality?
Are tests stale?
Does the team have sufficient tests?
Are any bottlenecks occurring?
31. Is the authoring of Test Cases on track?
Has the team defined Test Cases for all User Stories?
What are the proportions of Test Cases that are passing, failing, and blocked?
Do test failure metrics indicate a problem that requires further investigation?
What is the status of last night's build?
What are the most recent check-ins?
32. How quickly is the team resolving and closing bugs?
Is the team fixing bugs quickly enough to finish on time?
How many bugs is the team reporting, resolving, and closing per day?
Is the team resolving priority 1 bugs before priority 2 and 3 bugs?
Does any team member have a backlog of priority 1 bugs that warrant
redistribution?
33. How volatile is the code base?
How much of the code is the team testing?
How high is the quality of the builds?
Is the quality increasing, decreasing, or staying
constant?
Which builds succeeded?
Which builds have a significant number of changes
to the code?
Much of the content here is derived from published documentation on TFS reporting in the MSDN Library: http://msdn.microsoft.com/en-us/library/bb649552.aspx
Each tool or plug-in in Team Foundation uses a relational database in SQL Server 2008 to store the data used by the tool in its day-to-day operations. This relational database is often referred to as the operational store. The operational stores for Team Foundation include:Common structure databases (Tfs_Configuration)Team project collection databases (Tfs_Collection) You might also have operational stores created for third-party tools.Like most operational stores, the schema of the relational database is designed and optimized for the online transactional processing of data. As the tool or plug-in performs an activity, it writes the latest information to the operational store. Therefore, data in the operational store is constantly changing and being updated, and all data is current.
Because each tool or plug-in has its own schema requirements and data is stored in the operational store to optimize transactional processing, the purpose of the warehouse adapter is to put the operational data into a form usable by the data warehouse. The warehouse adapter is a managed assembly that extracts the data from the operational store, transforms the data to a standardized format compatible with the warehouse, and writes the transformed data into the warehouse relational database. There is a separate adapter for each operational data store. The warehouse adapter copies and transforms those data fields specified in either the basic warehouse configuration or in the process template used at the time a new team project is created. If you subsequently change the process template to add or delete which data fields are written to the data warehouse, these changes are detected the next time the adapter is run. The adapter runs periodically with a frequency set by the RunIntervalSeconds property. The default setting for the refresh frequency is two hours (7,200 seconds), so give careful consideration to the appropriate refresh frequency for your installation. For more information about changing the refresh frequency, see How to: Change the Refresh Frequency.It is important that data is not written from the relational database to the data cube while the relational database is itself being updated from the operational store. To avoid conflicts reading and writing data, the warehouse adapters that push and pull the data are synchronized. After the adapters have completed their calls, the cube is reprocessed.
Each tool describes its contribution to the data warehouse in an XML schema. The schema specifies the fields that are written to the relational database as dimensions, measures, and details. The schema is also mapped directly into the cube.The data in the warehouse are stored in a set of tables organized in a star schema. The central table of the star schema is called the fact table, and the related tables represent dimensions. Dimensions provide the means for disaggregating reports into smaller parts. A row in a fact table usually contains either the value of a measure or a foreign key reference to a dimension table. The row represents the current state of every item covered by the fact table. For example, the Work Item fact table has one row for every work item stored in Work Item operational store.A dimension table stores the set of values that exist for a given dimension. Dimensions may be shared between different fact tables and cubes, and they may be referenced by a single fact table or data cube. A Person dimension, for example, will be referenced by the Work Items fact table for Assigned To, Opened By, Resolved By, and Closed By properties, and it will be referenced by the Code Churn fact table for the Checked In By property.Measures are values taken from the operational data. For example, Total Churn is a measure that indicates the number of source code changes in the selected changesets. Count is a special measure in that it can be implicit, as long as there is one record for every item that is counted. The measures defined in a fact table form a measure group in the cube.For more information about the facts, dimensions, and measures in the data warehouse, see Perspectives and Measure Groups Provided in the Analysis Services Cube for Team System.
Fact tables are a good source of information for reports that show the current state of affairs. However, to report on trends for data that changes over time, you need to duplicate the same data for each of the time increments that you want to report on. For example, to report on daily trends for work items or test results, the warehouse needs to retain the state of every item for each day. This allows the data cube to aggregate the measures by day. The cube aggregates both data from the underlying star schema and time data into multidimensional structures.Each time the data cube is processed, the data stored in the star schemas in the relational database are pulled into the cube, aggregated, and stored. The data in the cube is aggregated so that high-level reports, which would otherwise require complex processing using the star schema, are simple select statements. The cube provides a central place to obtain data for reports without having to know the schema for each operational store and without having to access each store separately.
Report Designer is a component of Visual Studio that allows you to define the Team Foundation data warehouse as a data source and then design a report interactively. Report Designer provides tabbed windows for Data, Layout, and Preview, and you can add datasets to accommodate a new report design idea, or adjust report layout based on preview results. In addition to the Data, Layout, and Preview design surfaces, Report Designer provides query builders, an Expression editor, and wizards to help you place images or step you through the process of creating a simple report. For more information about using Report Designer, see Create, Customize, and Manage Reports for Visual Studio ALM.
Team Foundation integrates with Microsoft Excel to allow you to use Microsoft Excel to manage your project and produce reports. Microsoft Excel provides pivot tables and charts for viewing and analyzing multi-dimensional data. You can bind these pivot tables directly to the Team Foundation cube, so you can interact with the data in the cube. For more information about using Microsoft Excel for reporting, see Create and Manage Excel Reports for Visual Studio ALM.
You can generate several reports in Microsoft Excel that show current status and historical data based on the filter criteria that you specify in a flat-list work item query. This is useful to show the distribution of work items according to selected criteria or to view trends for the past several weeks. In addition, it is an effective way for you to quickly generate PivotTable and PivotChart reports that you can customize to support other report views. When you create an Excel report from a query, you can choose which reports to generate based on the variables that are used to filter the query and the criteria that you select. By using these methods, you can generate the following types of reports:Current reports: Pie charts that show the count of work items according to the filter criteria that are specified in the work item query. Trend reports: Line charts that show the distribution of work items over the past six weeks according to the filter criteria that are specified in the work item query. After the reports are generated, you can easily change the date range.Each report includes several worksheets, and each worksheet shows a PivotTable report and a PivotChart report that derives data from the SQL Server Analysis Services cube.
Team members can use the Bugs dashboard to determine whether they are managing the list of active Bugs according to established team goals and agile practices. By unit testing each increment of code before check-in, the team can reduce the overall number of bugs that the team must find. A team that focuses on being able to ship each increment of code removes defects incrementally and minimizes ongoing bugs.By using the Bugs dashboard, the team can answer the following questions:Is the number of active Bugs acceptable based on team goals? Is the team postponing too many Bugs?Is the team finding, fixing, and closing Bugs quickly enough to meet expectations and at a rate that matches previous development cycles? Is the team addressing high priority bugs before lower priority bugs?Does any team member need help in resolving bugs?
You can use the Code Coverage and Code Churn reports to answer the questions that are listed in the following table. Which builds succeeded? Which builds have a significant number of changes to the code? How often are builds succeeding?How volatile is the code base?How much of the code is the team testing?How high is the quality of the builds? Is the quality increasing, decreasing, or staying constant?
After the team has started to find and fix bugs, you can track the team's progress toward resolving and closing bugs by viewing the Bug Status report. This report shows the cumulative bug count based on the bug state, priority, and severity.
You can use the Bug Trends report to help track the rate at which your team is discovering and resolving bugs. This report shows a rolling or moving average of bugs being reported, resolved, and closed over time. When you manage a large team or a large number of bugs, you can monitor the Bug Trends report weekly to gain insight into how well the team is finding, resolving, and closing bugs.The Bug Trends report calculates a rolling average of the number of bugs that the team has opened, resolved, and closed based on the filters that you specify. The rolling average is based on the seven days before the date for which it is calculated. That is, the report averages the number of bugs in each state for each of the seven days before the date, and then the result is divided by seven.
As the team resolves and closes bugs, you can use the Reactivations report to determine how effectively the team is fixing bugs. Reactivations generally refer to bugs that have been resolved or closed prematurely and then reopened. The reactivation rate is also referred to as the fault feedback ratio.You can use the Reactivations report to show either bugs or user stories that have been reactivated. As a product owner, you might want to discuss acceptable rates of reactivation with the team. A low rate of reactivations (for example, less than 5%) might be acceptable depending on your team's goals. However, a high or increasing rate of reactivations indicates that the team might need to diagnose and fix systemic issues. The Reactivations report shows an area graph of the number of bugs or stories that are in a resolved state or that have been reactivated from the closed state.
The Build Quality Indicators report shows test coverage, code churn, and bug counts for a specified build definition. You can use this report to help determine how close portions of the code are to release quality. Ideally, test rates, bugs, and code churn would all produce the same picture, but they often do not. When you find a discrepancy, you can use the Bug Quality Indicators report to examine the details of a specific build and data series. Because this report combines test results, code coverage from testing, code churn, and bugs, you can view many perspectives at the same time.
The Build Success Over Time report provides a pictorial version of the Build Summary report. The Build Success Over Time report displays the status of the last build for each build category run for each day. You can use this report to help track the quality of the code that the team is checking in. In addition, for any day on which a build ran, you can view the Build Summary for that day.
The Build Summary lists builds and provides information about test results, test coverage, code churn, and quality notes for each build.The data that appears in the Build Summary report is derived from the data warehouse. The report presents a visual display of the percentage of tests that are passing, code that is being tested, and changes in code across several builds. You can review the results for both manual and automatic builds, in addition to the most recent builds and continuous or frequent builds. The report lists the most recent builds first and contains build results that were captured during the specified time interval for all builds that were run, subject to the filters that you specified for the report.At a glance, you can determine the success or failure of several build definitions for the time period under review.
After a team has worked on one or more iterations, also known as sprints, you can determine the rate of team progress by reviewing the Burndown and Burn Rate report. Burndown shows the trend of completed and remaining work over a specified time period. Burn rate provides calculations of the completed and required rate of work based on the specified time period. In addition, a chart shows the amount of completed and remaining work that is assigned to team members. You can view the Burndown and Burn Rate report based on hours worked or number of work items that have been resolved and closed.
After the team has estimated its tasks and begun work, you can use the Remaining Work report to track the team's progress and identify any problems in the flow of work. The Remaining Work report summarizes the data that was captured during the specified time interval for each task, user story, or bug based on the filter criteria that were specified for the report. The data is derived from the data warehouse.You can view this report in either the Hours of Work view or the Number of Work Items view. The first view displays the total number of hours of work for the specified time period and the team's progress toward completing that work. The second view displays the number of work items for the specified time period and the number of work items in each state. Each view provides an area graph that charts the progress of completed work against the total estimated work for the specified time duration.
After work has progressed on several iterations, also known as sprints, you can view the team progress by viewing the Status on All Iterations report. This report helps you track the team's performance over successive iterations. For each iteration that is defined for the product areas that you specify, this report displays the following information: Stories Closed: The number of user stories that have been closed. These values are derived from the current values specified for the iteration and the state of each user story.Progress (Hours): A two-bar numeric and visual representation that represents the values for Original Estimate (grey), Completed (green) and Remaining (light blue) based on the rollup of hours that are defined for all tasks. These values are derived from the current values that are specified for the iteration and the hours for each task. Bugs: A numeric value and visual representation for all bugs, grouped by their current states of Active (blue), Resolved (gold) and Closed (green). These values are derived from the current values that are specified for the iteration and the state of each bug.
The Stories Overview report lists all user stories, filtered by area and iteration and in order of importance.Work Progress% Hours Completed: A numeric value and visual representation that shows the percentage of completed work based on the rollup of baseline and completed hours for all tasks that are linked to the user story or its child stories.Hours Remaining: A numeric value for the rollup of all remaining hours for all tasks that are linked to the user story or its child stories.Test StatusTest Points: A numeric value that represents the number of pairings of test cases with test configurations in a specific test suite. For more information about test points, see Reporting on Testing Progress for Test Plans.Test Results: A numeric value and visual representation that shows the percentage of test cases, grouped according to the status of their most recent test run, where the options are Passed (green), Failed (red), or Not Run (black).Bugs: A numeric value and visual representation that shows the number of bugs that are linked to the test case or user story, where the options are Active (blue) and Resolved (gold). If a user story is linked to one or more child stories, the values represent a rollup of all bugs for the user story and its child stories.User Stories that Appear in the ReportThe Stories Overview report lists and highlights user stories according to the following criteria:Stories appear in order of their importance, based on their assigned ranking.Stories appear in bold type when they are in the active or resolved state. Stories appear in normal type when they are in the closed state. Stories appear in gray type when their assigned iteration or area is outside the filtered set, but they have tasks or child stories that are within the filtered set of iterations or product areas.
The Stories Progress report lists all user stories, filtered by product area and iteration in order of importance.This report displays the following information for each user story that appears in the report: Progress (% Completed): Numeric value that represents the percentage of completed work based on the rollup of baseline and completed hours for all tasks that are linked to the user story or its child stories.Hours Completed: A visual representation of the completed hours, displayed as a dark green bar.Recently Completed: A visual representation of those hours completed within the time interval specified for Recent (Calendar) Days, displayed as a light green bar.Hours Remaining: Rollup of all remaining hours for all tasks that are linked to the user story or its child stories.The Stories Progress report lists and highlights user stories according to the following criteria:Stories appear in order of their importance, based on their assigned ranking.Stories appear in bold type when they are in the active or resolved state. Stories appear in normal type when they are in the closed state. Stories appear in gray type when their assigned iteration or area is outside the filtered set but they have tasks or child stories that are within the filtered set of iterations or product areas.
The Requirements Progress report shows the status of completion as determined by the tasks that have been defined to implement the requirement.
The Requirements Overview report presents a snapshot of the work that has been performed for the filtered set of requirements to the current date.
By reviewing a release burndown report, you can understand how quickly your team has delivered backlog items and track how much work the team must still perform to complete a product release. A release burndown graph shows how much work remained at the start of each sprint in a release. The source of the raw data is your product backlog. Each sprint appears along the horizontal axis, and the vertical axis measures the effort that remained when each sprint started. The amount of estimated effort on the vertical axis is in whatever unit that your scrum team has decided to use (for example, story points or hours).
By reviewing a sprint burndown report, you can track how much work remains in a sprint backlog, understand how quickly your team has completed tasks, and predict when your team will achieve the goal or goals of the sprint.A sprint burndown report shows how much work remained at the end of specified intervals during a sprint. The source of the raw data is the sprint backlog. The horizontal axis shows days in a sprint, and the vertical axis measures the amount of work that remains to complete the tasks in the sprint. The work that remains is shown in hours. A sprint burndown graph displays the following pieces of data: The Ideal Trend line indicates an ideal situation in which the team burns down all of the effort that remains at a constant rate by the end of the sprint. The In Progress series shows how many hours remain for tasks that are marked as In Progress in a sprint. The To Do series shows how many hours remain for tasks that are marked as To Do in a sprint. Both the In Progress and the To Do series are drawn based on the actual progress of your team as it completes tasks.
Toward the end of an iteration, you can use the Unplanned Work report to determine how much work was added to the iteration that was not planned at the start of the iteration. You can view the unplanned work as measured by work items added, such as tasks, test cases, user stories, and bugs. Having unplanned work may be acceptable, especially if the team has scheduled a sufficient buffer for handling the load of unplanned work (for example, bugs). On the other hand, the unplanned work may represent a real problem if the team does not have the capacity to meet it and is forced to cut back on the planned work.The Unplanned Work report is useful when the team plans an iteration by identifying all work items that they intend to resolve or close during the course of the iteration. The work items that are assigned to the iteration by the plan completion date of the report are considered planned work. All work items that are added to the iteration after that date are identified as unplanned work.
The Test Case Readiness report provides an area graph that shows how many test cases are in the Design or Ready state over the time period that you specify. By reviewing this data, you can easily determine how quickly the team is designing test cases and making them ready for testing. When you create a test case, it is automatically set to the design state. After the team has reviewed and approved the test case, then a team member should change its state to Ready, which indicates that the test case is ready to be run.
The data that appears in the Test Plan Progress report is derived from the data warehouse and the test results that are generated when tests are run by using Microsoft Test Manager. The report presents an area graph that shows the most recent result of running any test in the specified test plans over time. For more information, see Running Tests.The horizontal axis shows days in a sprint or iteration, and the vertical axis shows test points. A test point is a pairing of a test case with a test configuration in a specific test suite. For more information about test points, see Reporting on Testing Progress for Test Plans.
http://msdn.microsoft.com/en-us/library/ms244710.aspxBy using the SQL Server Analysis Services cube for Visual Studio Team Foundation Server, you can generate reports of aggregated information about the data that is stored in team project collections. You can easily use this data to create PivotTable and PivotChart reports in Office Excel. You can drag the cube elements onto PivotTable or PivotChart reports to formulate questions and retrieve answers quickly. The cube is optimized to answer questions such as "How many bugs were active, resolved, and closed on each day of the project?"
A fact represents data that can be associated with multiple dimensions. This data may also be aggregated. Fact tables hold these values.Each data warehouse includes one or more fact tables. Central to a "star" or "snowflake" schema, a fact table captures the data that measures the team's operations. Fact tables usually contain large numbers of rows, especially when they contain one or more years of history for a large team project. A key characteristic of a fact table is that it contains numerical data (facts) that can be summarized to provide information about the history of the operation of the organization. Each fact table also includes a multipart index that contains, as foreign keys, the primary keys of related dimension tables. The related dimensions contain attributes of the fact records. Fact tables should not contain descriptive information or any data other than the numerical measurement fields and the index fields that relate the facts to corresponding entries in the dimension tables.For a list of the fact tables that are defined for the data warehouse, see Generate Reports Using the Relational Warehouse Database for Visual Studio ALM.
Each attribute is connected to a column in a corresponding dimension table in the data warehouse. Each dimension is associated with a set of attributes and potentially a set of hierarchies. Area and iteration paths are examples of hierarchies. Some work item dimension attributes are also stored as numeric and date filter values. When you use one of these dimension attributes in the rows or columns section, you can use these values to filter the report. For example, you can filter a report to show work items that were created after Oct 1, 2009, by using the value filter "System_CreatedDate is greater than Oct 1, 2009." You can also use the measure values to filter a report. For example, you can filter the report to show only work items that have more than two hours of work remaining by using the value filter "Remaining Work is greater than 2." For more information about value filters, see the following page on the Microsoft Web site: Filter Numbers in the Values Area.
Dimensions enable you to extract different views of data. Data values are associated with a set of dimensions that allow you to show aggregate results that are filtered using a specific set of dimension values.You can use dimensions to disaggregate the data and show more detail. For example, you can use the Date dimension in the rows or columns section of a PivotTable or PivotChart report to show a trend over time. You can also use dimensions to filter the report. Place a dimension or dimension attribute in the filter area, and then specify the values that you want to include in the report.Some dimensions are used in more than one measure group. For example, all measure groups share the Date, Team Project, Person, Area, and Iteration dimensions. The following illustration shows the dimensions in the cube.Dimensions are groups of attributes that are based on columns from tables or views in a data source view. Dimensions exist outside of a cube, can be used in multiple cubes, can be used multiple times in a single cube, and can be linked between Analysis Services instances. A dimension that exists outside of a cube is referred to as a database dimension, and an instance of a database dimension within a cube is referred to as a cube dimension.
In business terminology, a key performance indicator (KPI) is a quantifiable measurement for gauging business success.In Analysis Services, a KPI is a collection of calculations that are associated with a measure group in a cube and that are used to evaluate business success. Typically, these calculations are a combination of Multidimensional Expressions (MDX) expressions or calculated members. KPIs also have additional metadata that provides information about how client applications should display the results of the KPI's calculations.
Measures are values that correspond to columns in the corresponding fact table. Also, fields whose reportable attribute is set to Measure appear as measures in the cube.
Each measure group contains measures, such as Work Item Count, and dimensions, such as Date and Team Project. The measures are the numeric values that provide summaries at different levels of aggregation. You can use them in the Values section of a PivotTable or PivotChart report. The following illustration indicates the measure groups for Team Foundation.
By using perspectives, you can view portions of a cube, making it easier to focus on just the set of information that is of interest for creating a report. Note Perspectives are available only when your data warehouse for Visual Studio ALM is using SQL Server Enterprise Edition. Otherwise, you will see only a single perspective, the Team System cube.Each perspective provides a focused view of the data so that you do not have to scroll through all of the dimensions and measure groups that are defined for the whole cube. A perspective is a subset of the features and objects of a cube.