CS 206 Introduction to Computer Science II 02 / 02 / 2009 Instructor: Michael Eckmann
Michael Eckmann - Skidmore College - CS Spring 2009 Today’s Topics Questions/comments Solve 1 problem with three different algorithms. Analyze these algorithms. Linked lists (review?)
Michael Eckmann - Skidmore College - CS Spring 2009 Algorithm Analysis Let's consider 1 problem and 3 ways to solve it (using 3 different algorithms) and we'll analyze the running times of each. The Maximum contiguous subsequence problem: –Given an integer sequence A 1, A 2,..., A N, find (and identify the sequence corresponding to) the maximum value of j k=i A k. The maximum contiguous subsequence sum is zero if all are negative. Therefore, an empty sequence may be maximum. Example input: { -2, 11, -4, 13, -5, 2 } the answer is 20 and the sequence is { 11, -4, 13 } Another: { 1, -3, 4, -2, -1, 6 } the answer is 7, the sequence is { 4, -2, -1, 6 }
Michael Eckmann - Skidmore College - CS Spring 2009 Algorithm Analysis The simplest is an exhaustive search (brute force algorithm.) –that is, simply consider every possible subsequence and compute its sum, keep doing this and save the greatest –so, we set the maxSum = 0 (b/c it is at least this big) and we start at the first element and consider every subsequence that begins with the first element and sum them up... if any has a sum larger than maxSum, save this... –then start at second element and do the same... and so on until start at last element Advantages to this: not complex, easy to understand Disadvantages to this: slow Let's examine the algorithm. –decide what is a good thing to count –count that operation (in terms of the input size)
Michael Eckmann - Skidmore College - CS Spring 2009 Algorithm Analysis Let's consider this inefficient maximum continuous subsequence algorithm from page 171 in our text.
Michael Eckmann - Skidmore College - CS Spring 2009 Algorithm Analysis The exhaustive search has many unnecessary computations. Notice that j k=i A k = A j + j-1 k=i A k That is, if we know the sum of the first j-1 elements, then the sum of the first j elements is found just by adding in the jth element. Knowing that, the problem can be solved more efficiently by the algorithm that we are about to analyze. –We won't need to keep adding up a sequence from scratch
Michael Eckmann - Skidmore College - CS Spring 2009 Algorithm Analysis The second algorithm that we'll analyze uses the improvement just mentioned and the running time improves (goes down.)
Michael Eckmann - Skidmore College - CS Spring 2009 Algorithm Analysis A further improvement can come if we realize that if a subsequence has a negative sum, it will not be the first part of the maximum subsequence. –Why? Also, all contiguous subsequences bordering a maximum contiguous subsequence must have negative or 0 sums. –why?
Michael Eckmann - Skidmore College - CS Spring 2009 Algorithm Analysis A further improvement can come if we realize that if a subsequence has a negative sum, it will not be the first part of the maximum subsequence. –Why? –A negative value will only bring the total down (see Theorem 5.2 in textbook) Also, all contiguous subsequences bordering a maximum contiguous subsequence must have negative or 0 sums. –why? –If they were >0, they would be attached to the maximum sequence (thereby giving it a larger sum). Those two aren't too hard to see, but by themselves won't allow us to reduce the running time to below quadratic. Let's go a bit further.
Michael Eckmann - Skidmore College - CS Spring 2009 Algorithm Analysis There are some observations in our textbook which are then proved on page 175 in Theorem 5.3 of the textbook. This will allow us to process the sequence by just going through the sequence once from start to finish (and hence get linear time performance.) Specifically, while computing the sum of a subsequence, if at any time the sum becomes negative, we start considering sequences only starting at the next element. This is less intuitive than the other observations we made, but in my opinion not worth taking class time to cover in detail, so I leave it to you, if you wish to go over the observations in our textbook. But let's look at the linear algorithm and see if it is obvious that it is linear.
Michael Eckmann - Skidmore College - CS Spring 2009 Algorithm Analysis What's the point of that exercise: –1) get a feel for how to count how much work is being done in an algorithm –2) it is sometimes possible to get the running time of an algorithm down by exploiting facts about the problem. –3) it is good to think about such things in a course that mainly deals with data structures. Any guesses as to why I say this? –4) it is sometimes difficult to exploit some things about the problem to make a more efficient algorithm
Linked lists A linked list is a data structure where every node contains data and reference(s) to other node(s.) Let's look at drawings on the board. Let's figure out how to – Add to beginning of linked list – Add to end of linked list Need to find last node – Insert after a particular node – Delete a particular node Other options – Storing link to last node (tail) – Doubly linked lists (and their operations.) – Circular linked lists (where the “last” node refers to the “head”).
Linked lists A doubly linked list is a data structure where every node contains – data – a reference to previous node – a reference to next node The value of the next node for the last element of a linked list is null. The value of the previous node for the first element of a linked list is also null.
Linked lists In a singly linked list every node contains data and one reference to another node. –e.g. class Node { public AnyType data; // can have more data than one public Node next; } A Linked List is maintained by keeping a reference to the head of the list. From the head, we are able to get at all the other nodes in the list by following the next reference.
Linked lists class Node { public AnyType data; // can have more data than one public Node next; // when constructing a new node, we set its data and // set the next reference to null public Node(AnyType d) { data = d; next = null; }
Linked lists // All the code on the rest of the slides is outside of the Node class. Node n = new Node(somedata); Node head; head = n; // sets n to be the head of the linked list Node newnode = new Node(somedata2); // to add a newnode to the beginning of the list: newnode.next = head; head = newnode;
Linked lists Node newnode = new Node(somedata2); // to add a newnode to the end of the list: Node currnode; currnode = head; while (currnode != null) { savenode = currnode; currnode = currnode.next; } savenode.next = newnode;
Linked lists Insert after a particular node // insert newnode after findnode // head currnode = head; while (!((currnode.data).equals(findnode.data))) { currnode = currnode.next; } newnode.next = currnode.next; currnode.next = newnode;
Linked lists // delete findnode // head Node prevnode; Node currnode = head; while (!((currnode.data).equals(findnode.data))) { prevnode = currnode; currnode = currnode.next; } prevnode.next = currnode.next; currnode = null; // the data originally referred-to to be garbage collected // if we didn't have currnode = null; here, it would get garbage // collected once the currnode went out of scope.
Linked lists (chap. 4) What's deficient in singly linked lists?
Linked lists (chap. 4) Let's consider operations for a doubly linked list on the board. A doubly linked list is a data structure where every node has – data – a reference to previous node (prev) – a reference to next node (next) We'll also maintain a head node and a tail node. The value of the next node for the last (tail) element of a linked list is null. The value of the previous node for the first (head) element of a linked list is also null. When construct a new node, set the data and next and prev default to null