DPS921/Halt and Catch Fire

From CDOT Wiki
Revision as of 11:24, 8 December 2016 by Cpaul12 (talk | contribs) (Serial - Pi)
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Jump to: navigation, search

Halt and Catch Fire

Project: Parallelism with Go
Presentation: Slides

Group Members

  1. Colin Paul [1] Research etc.

Progress

Oct 31st:

  • Picked topic
  • Picked presentation date.
  • Created Wiki page

Nov 28th:

  • Presented topic

Parallel Programming in Go

What is Go?

  • The Go language is an open source project to make programmers more productive.
  • It is expressive, concise, clean, and efficient. Its concurrency mechanisms make it easy to write programs that get the most out of multi-core and networked machines, while its novel type system enables flexible and modular program construction. Go compiles quickly to machine code yet has the convenience of garbage collection and the power of run-time reflection. It's a fast, statically typed, compiled language that feels like a dynamically typed, interpreted language.

By what means does Go allow parallelism?

  • Go allows multi-core programming using concurrency methods, and enables the ability to parallelize.

TLDR;

concurrency vs parallelism

What is parallelism?

  • Programming as the simultaneous execution of (possibly related) computations.

What is concurrency?

  • Concurrency is the composition of independently executing computations.
  • It is a way to structure software, particularly as a way to write clean code that interacts well with the real world.
  • It is not parallelism, but it enables it.
  • If you have only one processor, your program can still be concurrent but it cannot be parallel.
  • On the other hand, a well-written concurrent program might run efficiently in parallel on a multi-processor.

Concurrency vs parallelism

  • Concurrency is about dealing with lots of things at once.
  • Parallelism is about doing lots of things at once.
  • Not the same, but related.
  • Concurrency is about structure, parallelism is about execution.
  • Concurrency provides a way to structure a solution to solve a problem that may (but not necessarily) be parallelizable.

Examples of things that use a concurrent model

  • I/O - Mouse, keyboard, display, and disk drivers.

Examples of things that use a parallel model

  • GPU - performing vector dot products.

Up and running with Go

Download and installation

Documentation

Playground

Demo: Concurrency and parallelism in Go

Monte Carlo Simulations

Monte Carlo Simulations
  • A probabilistic way to come up with the answer to a mathematical question by running a large number of simulations when you cannot get or want to double-check a closed-formed solution.
  • Use it to calculate the value of π (pi) - 3.1415926535

Method

  1. Draw a square, then inscribe a circle within it.
  2. Uniformly scatter some objects of uniform size (grains of rice or sand) over the square.
  3. Count the number of objects inside the circle and the total number of objects.
  4. The ratio of the two counts is an estimate of the ratio of the two areas, which is π/4. Multiply the result by 4 to estimate π.

Pseudo implementation

  • To make the simulations simple, we’ll use a unit square with sides of length 1.
  • That will make our final ratio of: (π*(1)2/4) / 12 = π/4
  • We just need to multiply by 4 to get π.

Programming implementation

Serial - Pi
package main

import (
  "fmt"
  "math/rand"
  "time"
)

//Function: Serial calculation of PI
func PI(samples int) float64 {
  var inside int = 0

  for i := 0; i < samples; i++ {
    x := rand.Float64()
    y := rand.Float64()
    if (x*x + y*y) < 1 {
      inside++
    }
  }

  ratio := float64(inside) / float64(samples)

  return ratio * 4
}

func init() {
    rand.Seed(time.Now().UnixNano())
}

func main() {
    fmt.Println("Our value of Pi after 100 runs:\t\t\t", PI(100))
    fmt.Println("Our value of Pi after 1,000 runs:\t\t", PI(1000))
    fmt.Println("Our value of Pi after 10,000 runs:\t\t", PI(10000))
    fmt.Println("Our value of Pi after 100,000 runs:\t\t", PI(100000))
    fmt.Println("Our value of Pi after 1,000,000 runs:\t\t", PI(1000000))
    fmt.Println("Our value of Pi after 10,000,000 runs:\t\t", PI(10000000))
    fmt.Println("Our value of Pi after 100,000,000 runs:\t\t", PI(100000000))
}
Serial execution CPU profile

As you can see only one core is working to compute the value of Pi - CPU8 @ 100%
Monte Carlo Simulations

Parallel - Pi
package main

import (
  "fmt"
  "math/rand"
  "runtime"
  "time"
)

// Struct: Stopwatch object
type StopWatch struct {
  start, stop time.Time
}

// Method: Calculate time delta
func (self *StopWatch) Milliseconds() uint32 {
  return uint32(self.stop.Sub(self.start) / time.Millisecond)
}

// Function: Start timer
func Start() time.Time {
  return time.Now()
}

// Function: Stop timer
func Stop(start time.Time) *StopWatch {
  watch := StopWatch{start: start, stop: time.Now()}
  return &watch
}

// Function: Serial calculation of PI
func PI(samples int) float64 {
  var inside int = 0

  r := rand.New(rand.NewSource(time.Now().UnixNano()))

  start := Start() // start timer

  for i := 0; i < samples; i++ {
    x := r.Float64()
    y := r.Float64()
    if (x*x + y*y) <= 1 {
      inside++
    }
  }

  ratio := float64(inside) / float64(samples)

  ratio = ratio * 4

  watch := Stop(start) // stop timer

  fmt.Printf("Serial - milliseconds elapsed:\t\t\t\t %v\n", watch.Milliseconds())

  return ratio
}

// Function: Concurrent calculation of PI
func MultiPI(samples int) float64 {

  cpus := runtime.NumCPU() // getting the numbre of CPUs

  fmt.Println("\nNumber of CPUs:\t\t\t\t\t\t", cpus)

  threadSamples := samples / cpus // splitting work among the CPUs

  results := make(chan float64, cpus)

  start := Start() // start timer

  for j := 0; j < cpus; j++ {
    go func() { // spawn goroutine
      var inside int
      r := rand.New(rand.NewSource(time.Now().UnixNano()))
      for i := 0; i < threadSamples; i++ {
        x, y := r.Float64(), r.Float64()
        if (x*x + y*y) <= 1 {
          inside++
        }
      }
      results <- float64(inside) / float64(threadSamples) * 4
    }()
  }

  var total float64

  // accumulate the results of reach channel
  for i := 0; i < cpus; i++ {
    total += <-results
  }

  total = total / float64(cpus)

  watch := Stop(start) // stop timer

  fmt.Printf("\nConcurrent - milliseconds elapsed:\t\t\t %v\n", watch.Milliseconds())

  return total
}

// get the number of CPUs available from the runtime
func init() {
    runtime.GOMAXPROCS(runtime.NumCPU())
}

func main() {
  fmt.Println("Running Monte Carlo simulations ...\n")
  fmt.Println("Our value of Pi after 1,000,000,000 runs:\t\t", PI(1000000000))
  fmt.Println("Our value of Pi after 1,000,000,000 runs (MT):\t\t", MultiPI(1000000000))
}
Parallel execution CPU profile

As you can see all eight cores are working to compute the value of Pi - CPU1-8 @ 100%
Monte Carlo Simulations

Results - Efficiency profile

Profiling the execution time at a computation hot spot for both the serial and parallel versions of the Monte Carlo Simulations program.
Monte Carlo Simulations
We can see that the parallel version gives us about a ~4x performance gain.

Results - Serial vs parallel performance

Evaluating the time it takes the serial and parallel versions of the program to run through the Monte Carlo Iterations for 1,000,000,000 (1 billion) iterations.
Monte Carlo Simulations
We can see that the parallel version gives us about a ~4x performance gain matching our benchmark results.