0 is a natural number. Or maybe it isn't.
When I was in high school (in Canada), my math teachers made a point of saying that the natural numbers are 1, 2, 3, ..., and that 0 is not natural. Then in university it seemed that everyone felt that other way, that the naturals should be 0, 1, 2, ... . The high school argument made more sense: 0 is unnatural because it represents the absence of things, and as evidenced by its relatively late historical appearance, it's a less natural number than 1, 2, 3, ... . But, it's just a definition, and it's more convenient to include 0 than to exclude it.
Of course 0 is a natural number. Burn in inquisition those who say it isn't. HaHaOnlySerious.
They additionally taught me in Canada that while the "natural" numbers were all positive integers starting at 1, the "whole" numbers were the same, with the inclusion of zero.
I am forever indebted to one of my early teachers who made it clear that the "common", "natural" or "usual" words were often misleading in mathematics.
Therefore, avoid the term "natural number" and the notation "N" unless either you are very sure of your audience or the distinction doesn't matter. Refer to "non-negative integers" and "positive integers". If you need a concise notation, you can write "Z" with a subscripted ">=0" or ">0".
The whole problem of confusing technical definitions with the everyday language equivalent is why NonEuclidean geometry got off to a slow start. In order to make sense of the resulting theorems, people had to make a broad leap of logic and rewrite the definition of a line; the practice (used in a lot of college TY math books) is that the plain word (such a line) means the everyday definition; while "line" or line means the technical one. Thus you can write "A line is defined as a circle on a sphere where the center of the circle has the same location as the center of the sphere." and it will make perfect sense. [ BelTorak ]
In describing EuclidOfAlexandria's description of the Greatest Common Divisor algorithm, DonKnuth remarked that for the ancient Greeks there was no number one.
I guess their thinking was that if you have only one thing, you don't have a "number" of them, now do you?
Nonnegative integers can be defined in terms of sets:
0 = {} (empty set) 1 = {{}} 2 = {{},{{}}} ... n = n-1 U {n-1}Can they really be? That is highly debatable. See for example Paul Benacerraf's famous paper "What numbers could not be". It's just common sense that numbers are not any particular abstract objects that happen to have all the same structural relations between them that numbers do (and then some extra ones as well -- 1 is not literally a subset (or a member) of 2, for example). I don't see what is so wrong with the traditional analysis of numbers that makes them out as properties of sets -- 2 being the property that all sets of 2 elements have in common, for example. Are the individuation critera for properties more difficult than those for sets, thus rendering the acceptance of properties unacceptable? They may be more difficult, but if you lost your keys in a big pile of keys, that doesn't justify looking for them in a desert landscape just because they would more clearly stand out there.
Unnatural... irrational... negative... imaginary.... Why this litany of pejoratives? Can't we all just get along? -(Pi)i. Maybe not: the problem is that we are all ultimately complex.
as evidenced by its relatively late historical appearance
Despite the late appearance of the numeral (which itself is debatable) & some silly musings by Greek philosophers (who also questioned whether one was a number), the number zero certainly didn't have a late appearance. Ever since people started counting & trading goods (if not before), people knew that if you took one away from one you'd be left with none.
Also see ChurchNumeral