Today was seminar day. I gave my talk about a brief introduction to Algorithm Analysis and I covered topics: a simplified computer architecture (with special attention to the role of the memory and of the CPU), a simple program for computing squares of integers and the concept of iteration, the accounting of steps of the program (which results in a linear polynomial in the number of squares printed), the fact that we can ignore constants when doing analysis of algorithms and asymptotic notation (O, \Omega, \Theta, o, \omega, \sim). Statement of the Prime Number Theorem in terms of asymptotic notation. Comments regarding difficulties of factoring integers, despite tests for primality being "easy".

It seems that I didn't scare people when talking such things, and that they felt like they understood what I said, which is good, since there are often conflicts between the language that Computer Science people (mainly the "practical people") use and the language that Mathematicians use. Since I am mostly in the middle, I think that I can talk with both without problems (but this means that I am not an specialist in anything, alas).