Re: sysadmin qualifications (Re: apt-get vs. aptitude)
Le 18.10.2013 19:36, Jerry Stuckle a écrit :
On 10/18/2013 1:10 PM, berenger.morel@neutralite.org wrote:
Le 18.10.2013 17:22, Jerry Stuckle a écrit :
On 10/17/2013 12:42 PM, berenger.morel@neutralite.org wrote:
Le 16.10.2013 17:51, Jerry Stuckle a écrit :
I only know few people who actually likes them :)
I liked them too, at a time, but since I can now use standard 
smart
pointers in C++, I tend to avoid them. I had so much troubles 
with
them,
so now I only use them for polymorphism and sometimes RTTI.
I hope that someday references will become usable in standard
containers... (I think they are not because of technical 
problems,
but I
do not know a lot about that. C++ is easy to learn, but hard to
master.)
Good design and code structure eliminates most pointer problems;
proper testing will get the rest.  Smart pointers are nice, but 
in
real time processing they are an additional overhead (and an 
unknown
one at that since you don't know the underlying libraries).
Depends on the smart pointer. shared_ptr indeed have a runtime 
cost,
since it maintains additional data, but unique_ptr does not, 
afaik, it
is made from pure templates, so only compilation-time cost.
You need to check your templates.  Templates generate code.  Code
needs resources to execute.  Otherwise there would be no difference
between a unique_ptr and a C pointer.
In practice, you can replace every occurrence of 
std::unique_ptr<int> by
int* in your code. It will still work, and have no bug. Except, of
course, that you will have to remove some ".get()", ".release()" and
things like that here and there.
You can not do the inverse transformation, because you can not copy
unique_ptr.
The only use of unique_ptr is to forbid some operations. The code it
generates is the same as you would have used around your raw 
pointers:
new, delete, swap, etc.
Of course, you can say that the simple fact of calling a method have 
an
overhead, but most of unique_ptr's stuff is inlined. Even without
speaking about compiler's optimizations.
Even inlined code requires resources to execute.  It is NOT as fast
as regular C pointers.
I did some testing, to be sure. With -O3, the code is exactly the same. 
Did not tried with -O1 and -O2. Without optimization, the 5 lines with 
pointers were half sized of those using unique_ptr. But I never ship 
softwares not optimized (the level depends on my needs, and usually I do 
not use -O3, though).
Plus, in an OS, there are applications. Kernels, drivers, and
applications.
Take windows, and say honestly that it does not contains 
applications?
explorer, mspaint, calc, msconfig, notepad, etc. Those are
applications,
nothing more, nothing less, and they are part of the OS. They 
simply
have to manage with the OS's API, as you will with any other
applications. Of course, you can use more and more layers 
between your
application the the OS's API, to stay in a pure windows 
environment,
there are (or were) for example MFC and .NET. To be more 
general, Qt,
wxWidgets, gtk are other tools.
mspaint, calc, notepad, etc. have nothing to do with the OS.  
They
are just applications shipped with the OS.  They run as user
applications, with no special privileges; they use standard
application interfaces to the OS, and are not required for any 
other
application to run.  And the fact they are written in C is 
immaterial.
So, what you name an OS is only drivers+kernel? If so, then ok. 
But some
people consider that it includes various other tools which does 
not
require hardware accesses. I spoke about graphical applications, 
and you
disagree. Matter of opinion, or maybe I did not used the good 
ones, I do
not know.
So, what about dpkg in debian? Is it a part of the OS? Is not it a 
ring
3 program? As for tar or shell?
Yes, the OS is what is required to access the hardware.  dpkg is an
application, as are tar and shell.
< snip >
Just because something is supplied with an OS does not mean it is
part of the OS.  Even DOS 1.0 came with some applications, like
command.com (the command line processor).
So, it was not a bad idea to ask what you name an OS. So, everything
which run in rings 0, 1 and 2 is part of the OS, but not softwares 
using
ring 3? Just for some confirmation.
Not necessarily.  There are parts of the OS which run at ring 3, 
also.
What's important is not what ring it's running at - it's is the code
required to access the hardware on the machine?
I disagree, but it is not important, since at least now I can use 
the
word in the same meaning as you, which is far more important.
But all of this have nothing related to the need of 
understanding
basics
of what you use when doing a program. Not understanding how a
resources
you acquired works in its big lines, imply that you will not be
able to
manage it correctly by yourself. It is valid for RAM memory, but 
also
for CPU, network sockets, etc.
Do you know how the SQL database you're using works?
No, but I do understand why comparing text is slower than integers 
on
x86 computers. Because I know that an int can be stored into one 
word,
which can be compared with only one instruction, while the text 
will
imply to compare more than one word, which is indeed slower. And 
it can
even become worse when the text is not an ascii one.
So I can use that understanding to know why I often avoid to use 
text as
keys. But it happens that sometimes the more problematic cost is 
not the
speed but the memory, and so sometimes I'll use text as keys 
anyway.
Knowing what is the word's size of the SQL server is not needed to 
make
things work, but it is helps to make it working faster. Instead of
requiring to buy more hardware.
First of all, there is no difference between comparing ASCII text 
and
non-ASCII text, if case-sensitivity is observed.
Character's size, in bits. ASCII uses 7 bits, E-ASCII uses 8, UTF8 = 
8,
UTF16 = 16, etc. It have an impact, for both memory, bandwidth and
instruction sets used.
But ASCII, even if it only uses 7 bits, is stored in an 8 bit byte.
A 4 byte ASCII character will take up exactly the same amount of room
as a 32 bit integer.  And comparison can use exactly the same machine
language instructions for both.
Be fair. Compare the max number you can represent with a uint32_t with 
the max you can represent with a char*.
For int8_t and char:
256 values versus 10 if we limit ourselves to numbers. If we limit 
ourselves to printable ASCII characters, ok, we can have... a little 
more than 100, but what a strange numeric base it will be...
The exact same set
of machine language instructions is generated.  However, if you are
doing a case-insensitive comparison, ASCII is definitely slower.
And saying "comparing text is slower than integers" is completely
wrong.  For instance, a CHAR(4) field can be compared just as 
quickly
as an INT field, and CHAR(2) may in fact be faster, depending on 
many
factors.
It is partially wrong. Comparing a text of 6 characters will be 
slower
than comparing short.
6 characters: "-12345" and you have the same data on only 2 bytes.
I didn't say 6 characters.  I SPECIFICALLY said 4 characters - one
case where your "strings take longer to compare than integers) is
wrong.
With you char[4], you can only go to 9999, even a short can represent 
more, and so, compute (including comparison) faster.
Now, you say I'm completely wrong, when you take 1 case where I am 
wrong. I agreed with you that I was partially, but by saying that 
comparing text is slower than comparing integers, I did not specifically 
same binary size. Plus, in the SQL context, I thought it was obvious 
that I was referring to the fact that sometimes integers are good 
replacements for primary keys.
But, fine. You said, without taking uppercase/lowercase, that's the 
same?
So, how would you sort texts which includes, for example: éèê ?
Those chars are greater than 'f' if you take only their numerical 
values. But then, results would be... just wrong. They should, at least 
in French, be considered equal with the 'e'. That's what users would 
expect. So the comparison of texts will need replacement and then 
comparison, for the simpler (and the worse, since it does not sort the 
accentuated characters themselves plus it loses informations) solution. 
And we have plenty of characters like that here, that we use on lot of 
words (and names, indeed).
Texts are directly linked with localization, so you can not simply 
compare them as if they were integers. It only works with English ( or 
maybe there is another language which uses only 26 letters without 
accents, but I do not know about it). The old trick to consider chars as 
bytes should not be taught longer in my opinion. Students could use it 
real situations where there are accentuated characters, and cause bugs 
for nothing.
Plus, CHAR(4) is not necessarily coded on 4 bytes. Characters and 
bytes
are different notions.
In any current database, CHAR(4) for ASCII data is encoded in 4
bytes. Please show where that is not the case.
Ok, you got me one the char(4) stuff. All that time I was thinking 
about text, but forgot to be explicit enough. I thought I mentioned 
unicode somewhere... but I'm too lazy to check it.
As a side note, I'm surprised. CHAR does not seems to be in the 
standard:
http://www.google.com/url?sa=t&rct=j&q=&esrc=s&source=web&cd=1&ved=0CCwQFjAA&url=http%3A%2F%2Fjtc1sc32.org%2Fdoc%2FN1951-2000%2F32N1964T-text_for_ballot-FCD_9075-2.pdf&ei=6JBlUoPXNoa00QWCrIGYAw&usg=AFQjCNHiJl_XShEUGPmObfmrji81RtDVNg&sig2=1vIOHIp64_oLVO8rIuMjIA&bvm=bv.54934254,d.d2k
But if an extra 4 byte key is going to cause you memory problems,
you're hardware is already undersized.
Or your program could be too hungry, because you did not know that 
you
have a limited hardware.
As I said - your hardware is already undersized.  If adding 4 bytes
to a row is going to cause problems now, you'll have even greater
problems later.
Smaller stuff is often better. It is not a problem of hardware being 
undersized.
And here, we are not in the simple efficiency, but to something 
which
can make an application completely unusable, with "random" errors.
Not at all.  Again, it's a matter of understanding the language you
are using.  Different languages have different limitations.
So it must be that C's limitations are not fixed enough, because 
size
types can vary according to the hardware (and/or compiler).
Sure.  And you need to understand those limitations.
Indeed. And those are dependent on hardware and compiler, for the C and 
C++ languages at least.
And BTW - even back in the days if 16 bit PC's, C compilers still
used 32 bit ints.
I can remember having used borland TurboC (can't remember the version, 
I do not even remember if it was able to support C++), and it's "int" 
type was a short. 16 bits. It is when I switched to another (more recent 
one) compiler that I stopped using int, for this exact reason. And when 
I discovered stdint.h, I simply immediately loved it. No more surprises, 
no implicitly defined sizes.
So, ok, if you can find a job when you have every single low level
feature you will require through high level functions/objects, 
having
knowledge of on what you are sit on is useless. Maybe I am wrong 
because
I actually am interested by knowing what is behind the screen, and 
not
only by the screen itself. But still, if you only know about your 
own
stuff, and the man who will deploy it only knows about his own 
stuff,
won't you need a 3rd person to allow you to communicate? Which 
imply
loss of time.
No, it's called a DESIGN - which, amongst other things, defines the
communications method between the two.  I've been on projects with 
up
to 100 or so programmers working on various pieces.  But everything
worked because it was properly designed from the outset and every
programmer knew how his/her part fit into the whole programs.
I do not think that most programmers work in teams of hundreds of
people. But I may be wrong. I do not know.
I didn't say most did.  I DID say they exist, for large projects.
I never said you said all did. I simply said I do not know the average 
size of programmer's teams around the world.
I think, but it's pure guessing based on my small experience, that IT 
services with more than 30 persons in R&D are not the most common 
situations, and if I am right, then programmers... no, people, have to 
be able to interact with other teams which are doing different things.
And to interact, you need to be able to speak the same language.
At least, that's what I have seen in my experiences. But, indeed, my cv 
is far less impressive than yours, and I would never use it to prove 
that I am true.
(note: we were exactly 6 programmers. There were 3 sysadmins, 1 lead 
project, and 3 others were for support.)
In my last job, when we had something to release, we usually 
talked
directly with the people who had then to deploy it, to explain 
them some
requirements and consequences, that were not directly our 
programmer's
job. Indeed, I was not employed by microsoft, google or IBM, but 
very
far from that, we were less than 10 dev.
But now, are most programmers paid by societies with hundreds of
programmers?
In the jobs I've had, the programmers have never had to talk to the
deployers.  Both were given the design details; the programmers 
wrote
to the design and the deployers generated the necessary scripts do
deploy what the design indicated.  When the programmers were done, 
the
deployers were ready to install.
Maybe you worked only in big structures, or maybe this one was doing
things wrong. But the IT team was quite small, if we only consider
sysadmins, dev, project leads and few other roles. Less than 15 
persons.
Even 15 person teams can do it right.  Unfortunately, too many
companies (both big and small) won't do it right.  A major reason why
there are so many bugs out there.  Also a major reason why projects 
go
over time and over budget.
Sorry, but I do not see what's wrong with having communication between 
various actors of a project.
I have followed various classes of various good ways to manage 
projects, stuff about ITIL for example, and I can not remember having 
learn that part of the process was wrong in my last job. Other were, and 
not only a little, but I do not think having learn that something was 
wrong in that part. But that was only lessons, and I never read the 
whole ITIL stuff.
Reply to: