aboutsummaryrefslogtreecommitdiff
path: root/doc/FAQ_DEV
diff options
context:
space:
mode:
authorMagnus Hagander <magnus@hagander.net>2008-04-22 09:26:41 +0000
committerMagnus Hagander <magnus@hagander.net>2008-04-22 09:26:41 +0000
commita18d6de197aa9dc1630e3381399df1e9c197bba4 (patch)
treedb66ed889ed60ca520b3a99d5bfe9216daf30753 /doc/FAQ_DEV
parentc68182e2cc7e67b13e1bf7866c7054a085e9bc4a (diff)
downloadpostgresql-a18d6de197aa9dc1630e3381399df1e9c197bba4.tar.gz
postgresql-a18d6de197aa9dc1630e3381399df1e9c197bba4.zip
Replace developer FAQ with a reference to the wiki, which is where
it now lives (per discussion). Leave the other FAQs alone for now.
Diffstat (limited to 'doc/FAQ_DEV')
-rw-r--r--doc/FAQ_DEV843
1 files changed, 2 insertions, 841 deletions
diff --git a/doc/FAQ_DEV b/doc/FAQ_DEV
index e3816b2682f..3f2070aff80 100644
--- a/doc/FAQ_DEV
+++ b/doc/FAQ_DEV
@@ -1,842 +1,3 @@
+The developer FAQ can be found on the PostgreSQL wiki:
- Developer's Frequently Asked Questions (FAQ) for PostgreSQL
-
- Last updated: Wed Aug 22 20:10:01 EDT 2007
-
- Current maintainer: Bruce Momjian (bruce@momjian.us)
-
- The most recent version of this document can be viewed at
- http://www.postgresql.org/docs/faqs.FAQ_DEV.html.
- _________________________________________________________________
-
-General Questions
-
- 1.1) How do I get involved in PostgreSQL development?
- 1.2) What development environment is required to develop code?
- 1.3) What areas need work?
- 1.4) What do I do after choosing an item to work on?
- 1.5) I have developed a patch, what next?
- 1.6) How is a patch reviewed?
- 1.7) Where can I learn more about the code?
- 1.8) How do I download/update the current source tree?
- 1.9) How do I test my changes?
- 1.10) What tools are available for developers?
- 1.11) What books are good for developers?
- 1.12) What is configure all about?
- 1.13) How do I add a new port?
- 1.14) Why don't you use threads, raw devices, async-I/O, <insert your
- favorite wizz-bang feature here>?
- 1.15) How are RPM's packaged?
- 1.16) How are CVS branches handled?
- 1.17) Where can I get a copy of the SQL standards?
- 1.18) Where can I get technical assistance?
- 1.19) How do I get involved in PostgreSQL web site development?
- 1.20) Why haven't you replaced CVS with SVN, Git, Monotone, VSS,
- <insert your favorite SCM system here>?
-
-Technical Questions
-
- 2.1) How do I efficiently access information in tables from the
- backend code?
- 2.2) Why are table, column, type, function, view names sometimes
- referenced as Name or NameData, and sometimes as char *?
- 2.3) Why do we use Node and List to make data structures?
- 2.4) I just added a field to a structure. What else should I do?
- 2.5) Why do we use palloc() and pfree() to allocate memory?
- 2.6) What is ereport()?
- 2.7) What is CommandCounterIncrement()?
- 2.8) What debugging features are available?
- _________________________________________________________________
-
-General Questions
-
- 1.1) How do I get involved in PostgreSQL development?
-
- Download the code and have a look around. See 1.8.
-
- Subscribe to and read the pgsql-hackers mailing list (often termed
- 'hackers'). This is where the major contributors and core members of
- the project discuss development.
-
- 1.2) What development environment is required to develop code?
-
- PostgreSQL is developed mostly in the C programming language. It also
- makes use of Yacc and Lex.
-
- The source code is targeted at most of the popular Unix platforms and
- the Windows environment (XP, Windows 2000, and up).
-
- Most developers make use of the open source development tool chain. If
- you have contributed to open source software before, you will probably
- be familiar with these tools. They include: GCC (http://gcc.gnu.org,
- GDB (www.gnu.org/software/gdb/gdb.html), autoconf
- (www.gnu.org/software/autoconf/) AND GNU make
- (www.gnu.org/software/make/make.html.
-
- Developers using this tool chain on Windows make use of MingW (see
- http://www.mingw.org/).
-
- Some developers use compilers from other software vendors with mixed
- results.
-
- Developers who regularly rebuild the source often pass the
- --enable-depend flag to configure. The result is that when you make a
- modification to a C header file, all files depend upon that file are
- also rebuilt.
-
- src/Makefile.custom can be used to set environment variables, like
- CUSTOM_COPT, that are used for every compile.
-
- 1.3) What areas need work?
-
- Outstanding features are detailed in the TODO list. This is located in
- doc/TODO in the source distribution or at
- http://www.postgresql.org/docs/faqs.TODO.html.
-
- You can learn more about these features by consulting the archives,
- the SQL standards and the recommend texts (see 1.11).
-
- 1.4) What do I do after choosing an item to work on?
-
- Send an email to pgsql-hackers with a proposal for what you want to do
- (assuming your contribution is not trivial). Working in isolation is
- not advisable because others might be working on the same TODO item,
- or you might have misunderstood the TODO item. In the email, discuss
- both the internal implementation method you plan to use, and any
- user-visible changes (new syntax, etc). For complex patches, it is
- important to get community feeback on your proposal before starting
- work. Failure to do so might mean your patch is rejected. If your work
- is being sponsored by a company, read this article for tips on being
- more effective.
-
- A web site is maintained for patches awaiting review,
- http://momjian.postgresql.org/cgi-bin/pgpatches, and those that are
- being kept for the next release,
- http://momjian.postgresql.org/cgi-bin/pgpatches_hold.
-
- 1.5) I have developed a patch, what next?
-
- You will need to submit the patch to pgsql-patches@postgresql.org. It
- will be reviewed by other contributors to the project and will be
- either accepted or sent back for further work. To help ensure your
- patch is reviewed and committed in a timely fashion, please try to
- make sure your submission conforms to the following guidelines:
- 1. Ensure that your patch is generated against the most recent
- version of the code, which for developers is CVS HEAD. For more on
- branches in PostgreSQL, see 1.16.
- 2. Try to make your patch as readable as possible by following the
- project's code-layout conventions. This makes it easier for the
- reviewer, and there's no point in trying to layout things
- differently than pgindent. Also avoid unnecessary whitespace
- changes because they just distract the reviewer, and formatting
- changes will be removed by the next run of pgindent.
- 3. The patch should be generated in contextual diff format (diff -c
- and should be applicable from the root directory. If you are
- unfamiliar with this, you might find the script
- src/tools/make_diff/difforig useful. (Unified diffs are only
- preferable if the file changes are single-line changes and do not
- rely on surrounding lines.)
- 4. PostgreSQL is licensed under a BSD license. By posting a patch to
- the public PostgreSQL mailling lists, you are giving the
- PostgreSQL Global Development Group the non-revokable right to
- distribute your patch under the BSD license.
- 5. Confirm that your changes can pass the regression tests. If your
- changes are port specific, please list the ports you have tested
- it on.
- 6. If you are adding a new feature, confirm that it has been tested
- thoroughly. Try to test the feature in all conceivable scenarios.
- 7. New feature patches should also be accompanied by documentation
- patches. If you need help checking the SQL standard, see 1.17.
- 8. Provide an implementation overview, preferably in code comments.
- Following the surrounding code commenting style is usually a good
- approach (also see
- http://www.ibm.com/developerworks/linux/library/l-clear-code/?ca=d
- gr-FClnxw01linuxcodetips).
- 9. If it is a performance patch, please provide confirming test
- results to show the benefit of your patch. It is OK to post
- patches without this information, though the patch will not be
- applied until somebody has tested the patch and found a
- significant performance improvement.
-
- Even if you pass all of the above, the patch might still be rejected
- for other reasons. Please be prepared to listen to comments and make
- modifications.
-
- You will be notified via email when the patch is applied, and your
- name will appear in the next version of the release notes.
-
- 1.6) How is a patch reviewed?
-
- Patch committers check several things before applying a patch:
- * Patch follows the SQL standard or community agreed-upon behavior
- * Style merges seamlessly into the surrounding code
- * Written as simply and efficiently as possible
- * Uses the available PostgreSQL subsystems properly
- * Contains sufficient comments
- * Contains code that works on all supported operating systems
- * Has proper documentation
- * Passes all regression tests, and if needed, adds new ones
- * Behaves as expected, even under unusual cirumstances
- * Contains no reliability risks
- * Does not overly complicate the source code
- * If performance-related, has a measureable performance benefit
- * Is of sufficient usefulness to the average PostgreSQL user
- * Follows existing PostgreSQL coding standards
-
- 1.7) Where can I learn more about the code?
-
- Other than documentation in the source tree itself, you can find some
- papers/presentations discussing the code at
- http://www.postgresql.org/developer. An excellent presentation is at
- http://neilconway.org/talks/hacking/
-
- 1.8) How do I download/update the current source tree?
-
- There are several ways to obtain the source tree. Occasional
- developers can just get the most recent source tree snapshot from
- ftp://ftp.postgresql.org.
-
- Regular developers might want to take advantage of anonymous access to
- our source code management system. The source tree is currently hosted
- in CVS. For details of how to obtain the source from CVS see
- http://developer.postgresql.org/docs/postgres/cvs.html.
-
- 1.9) How do I test my changes?
-
- Basic system testing
-
- The easiest way to test your code is to ensure that it builds against
- the latest version of the code and that it does not generate compiler
- warnings.
-
- It is worth advised that you pass --enable-cassert to configure. This
- will turn on assertions with in the source which will often show us
- bugs because they cause data corruption of segmentation violations.
- This generally makes debugging much easier.
-
- Then, perform run time testing via psql.
-
- Regression test suite
-
- The next step is to test your changes against the existing regression
- test suite. To do this, issue "make check" in the root directory of
- the source tree. If any tests failure, investigate.
-
- If you've deliberately changed existing behavior, this change might
- cause a regression test failure but not any actual regression. If so,
- you should also patch the regression test suite.
-
- Other run time testing
-
- Some developers make use of tools such as valgrind
- (http://valgrind.kde.org) for memory testing, gprof (which comes with
- the GNU binutils suite) and oprofile
- (http://oprofile.sourceforge.net/) for profiling and other related
- tools.
-
- What about unit testing, static analysis, model checking...?
-
- There have been a number of discussions about other testing frameworks
- and some developers are exploring these ideas.
-
- Keep in mind the Makefiles do not have the proper dependencies for
- include files. You have to do a make clean and then another make. If
- you are using GCC you can use the --enable-depend option of configure
- to have the compiler compute the dependencies automatically.
-
- 1.10) What tools are available for developers?
-
- First, all the files in the src/tools directory are designed for
- developers.
- RELEASE_CHANGES changes we have to make for each release
- backend description/flowchart of the backend directories
- ccsym find standard defines made by your compiler
- copyright fixes copyright notices
-
- entab converts spaces to tabs, used by pgindent
- find_static finds functions that could be made static
- find_typedef finds typedefs in the source code
- find_badmacros finds macros that use braces incorrectly
- fsync a script to provide information about the cost of cache
- syncing system calls
- make_ctags make vi 'tags' file in each directory
- make_diff make *.orig and diffs of source
- make_etags make emacs 'etags' files
- make_keywords make comparison of our keywords and SQL'92
- make_mkid make mkid ID files
- pgcvslog used to generate a list of changes for each release
- pginclude scripts for adding/removing include files
- pgindent indents source files
- pgtest a semi-automated build system
- thread a thread testing script
-
- In src/include/catalog:
- unused_oids a script which generates unused OIDs for use in system
- catalogs
- duplicate_oids finds duplicate OIDs in system catalog definitions
-
- If you point your browser at the tools/backend/index.html file, you
- will see few paragraphs describing the data flow, the backend
- components in a flow chart, and a description of the shared memory
- area. You can click on any flowchart box to see a description. If you
- then click on the directory name, you will be taken to the source
- directory, to browse the actual source code behind it. We also have
- several README files in some source directories to describe the
- function of the module. The browser will display these when you enter
- the directory also. The tools/backend directory is also contained on
- our web page under the title How PostgreSQL Processes a Query.
-
- Second, you really should have an editor that can handle tags, so you
- can tag a function call to see the function definition, and then tag
- inside that function to see an even lower-level function, and then
- back out twice to return to the original function. Most editors
- support this via tags or etags files.
-
- Third, you need to get id-utils from ftp://ftp.gnu.org/gnu/id-utils/
-
- By running tools/make_mkid, an archive of source symbols can be
- created that can be rapidly queried.
-
- Some developers make use of cscope, which can be found at
- http://cscope.sf.net/. Others use glimpse, which can be found at
- http://webglimpse.net/.
-
- tools/make_diff has tools to create patch diff files that can be
- applied to the distribution. This produces context diffs, which is our
- preferred format.
-
- Our standard format BSD style, with each level of code indented one
- tab, where each tab is four spaces. You will need to set your editor
- or file viewer to display tabs as four spaces:
- vi in ~/.exrc:
- set tabstop=4
- set sw=4
- more:
- more -x4
- less:
- less -x4
-
- The tools/editors directory of the latest sources contains sample
- settings that can be used with the emacs, xemacs and vim editors, that
- assist in keeping to PostgreSQL coding standards.
-
- pgindent will the format code by specifying flags to your operating
- system's utility indent. This article describes the value of a
- consistent coding style.
-
- pgindent is run on all source files just before each beta test period.
- It auto-formats all source files to make them consistent. Comment
- blocks that need specific line breaks should be formatted as block
- comments, where the comment starts as /*------. These comments will
- not be reformatted in any way.
-
- pginclude contains scripts used to add needed #include's to include
- files, and removed unneeded #include's.
-
- When adding system types, you will need to assign oids to them. There
- is also a script called unused_oids in pgsql/src/include/catalog that
- shows the unused oids.
-
- 1.11) What books are good for developers?
-
- There are five good books:
- * An Introduction to Database Systems, by C.J. Date, Addison, Wesley
- * A Guide to the SQL Standard, by C.J. Date, et. al, Addison, Wesley
- * Fundamentals of Database Systems, by Elmasri and Navathe
- * Transaction Processing, by Jim Gray, Morgan, Kaufmann
- * Transactional Information Systems by Gerhard Weikum, Kaufmann
-
- 1.12) What is configure all about?
-
- The files configure and configure.in are part of the GNU autoconf
- package. Configure allows us to test for various capabilities of the
- OS, and to set variables that can then be tested in C programs and
- Makefiles. Autoconf is installed on the PostgreSQL main server. To add
- options to configure, edit configure.in, and then run autoconf to
- generate configure.
-
- When configure is run by the user, it tests various OS capabilities,
- stores those in config.status and config.cache, and modifies a list of
- *.in files. For example, if there exists a Makefile.in, configure
- generates a Makefile that contains substitutions for all @var@
- parameters found by configure.
-
- When you need to edit files, make sure you don't waste time modifying
- files generated by configure. Edit the *.in file, and re-run configure
- to recreate the needed file. If you run make distclean from the
- top-level source directory, all files derived by configure are
- removed, so you see only the file contained in the source
- distribution.
-
- 1.13) How do I add a new port?
-
- There are a variety of places that need to be modified to add a new
- port. First, start in the src/template directory. Add an appropriate
- entry for your OS. Also, use src/config.guess to add your OS to
- src/template/.similar. You shouldn't match the OS version exactly. The
- configure test will look for an exact OS version number, and if not
- found, find a match without version number. Edit src/configure.in to
- add your new OS. (See configure item above.) You will need to run
- autoconf, or patch src/configure too.
-
- Then, check src/include/port and add your new OS file, with
- appropriate values. Hopefully, there is already locking code in
- src/include/storage/s_lock.h for your CPU. There is also a
- src/makefiles directory for port-specific Makefile handling. There is
- a backend/port directory if you need special files for your OS.
-
- 1.14) Why don't you use threads, raw devices, async-I/O, <insert your
- favorite wizz-bang feature here>?
-
- There is always a temptation to use the newest operating system
- features as soon as they arrive. We resist that temptation.
-
- First, we support 15+ operating systems, so any new feature has to be
- well established before we will consider it. Second, most new
- wizz-bang features don't provide dramatic improvements. Third, they
- usually have some downside, such as decreased reliability or
- additional code required. Therefore, we don't rush to use new features
- but rather wait for the feature to be established, then ask for
- testing to show that a measurable improvement is possible.
-
- As an example, threads are not currently used in the backend code
- because:
- * Historically, threads were unsupported and buggy.
- * An error in one backend can corrupt other backends.
- * Speed improvements using threads are small compared to the
- remaining backend startup time.
- * The backend code would be more complex.
-
- So, we are not ignorant of new features. It is just that we are
- cautious about their adoption. The TODO list often contains links to
- discussions showing our reasoning in these areas.
-
- 1.15) How are RPMs packaged?
-
- This was written by Lamar Owen and Devrim Gündüz:
-
- 2006-10-16
-
- As to how the RPMs are built -- to answer that question sanely
- requires us to know how much experience you have with the whole RPM
- paradigm. 'How is the RPM built?' is a multifaceted question. The
- obvious simple answer is that we maintain:
- 1. A set of patches to make certain portions of the source tree
- 'behave' in the different environment of the RPMset;
- 2. The initscript;
- 3. Any other ancillary scripts and files;
- 4. A README.rpm-dist document that tries to adequately document both
- the differences between the RPM build and the WHY of the
- differences, as well as useful RPM environment operations (like,
- using syslog, upgrading, getting postmaster to start at OS boot,
- etc);
- 5. The spec file that throws it all together. This is not a trivial
- undertaking in a package of this size.
-
- PGDG RPM Maintainer builds the SRPM and announces the SRPM to the
- pgsqlrpms-hackers list. This is a list where package builders are
- subscribed. Then, the builders download the SRPM and rebuild it on
- their machines.
-
- We try to build on as many different canonical distributions as we
- can. Currently we are able to build on Red Hat Linux 9, RHEL 3 and
- above, and all Fedora Core Linux releases.
-
- To test the binaries, we install them on our local machines and run
- regression tests. If the package builders uses postgres user to build
- the rpms, then it is possible to run regression tests during RPM
- builds.
-
- Once the build passes these tests, the binary RPMs are sent back to
- PGDG RPM Maintainer and they are pushed to main FTP site, followed by
- a release announcement to pgsqlrpms-* lists, pgsql-general and
- pgsql-announce lists.
-
- You will notice we said 'canonical' distributions above. That simply
- means that the machine is as stock 'out of the box' as practical --
- that is, everything (except select few programs) on these boxen are
- installed by RPM; only official Red Hat released RPMs are used (except
- in unusual circumstances involving software that will not alter the
- build -- for example, installing a newer non-RedHat version of the Dia
- diagramming package is OK -- installing Python 2.1 on the box that has
- Python 1.5.2 installed is not, as that alters the PostgreSQL build).
- The RPM as uploaded is built to as close to out-of-the-box pristine as
- is possible. Only the standard released 'official to that release'
- compiler is used -- and only the standard official kernel is used as
- well.
-
- PGDG RPM Building Project does not build RPMs for Mandrake .
-
- We usually have only one SRPM for all platforms. This is because of
- our limited resources. However, on some cases, we may distribute
- different SRPMs for different platforms, depending on possible
- compilation problems, especially on older distros.
-
- Please note that this is a volunteered job -- We are doing our best to
- keep packages up to date. We, at least, provide SRPMs for all
- platforms. For example, if you do not find a RHEL 4 x86_64 RPM in our
- FTP site, it means that we do not have a RHEL 4 x86_64 server around.
- If you have one and want to help us, please do not hesitate to build
- rpms and send to us :-)
- http://pgfoundry.org/docman/view.php/1000048/98/PostgreSQL-RPM-Install
- ation-PGDG.pdf has some information about building binary RPMs using
- an SRPM.
-
- PGDG RPM Building Project is a hosted on pgFoundry :
- http://pgfoundry.org/projects/pgsqlrpms. We are an open community,
- except one point : Our pgsqlrpms-hackers list is open to package
- builders only. Still, its archives are visible to public. We use a CVS
- server to save the work we have done so far. This includes spec files
- and patches; as well as documents.
-
- As to why all these files aren't part of the source tree, well, unless
- there was a large cry for it to happen, we don't believe it should.
-
- 1.16) How are CVS branches managed?
-
- This was written by Tom Lane:
-
- 2001-05-07
-
- If you just do basic "cvs checkout", "cvs update", "cvs commit", then
- you'll always be dealing with the HEAD version of the files in CVS.
- That's what you want for development, but if you need to patch past
- stable releases then you have to be able to access and update the
- "branch" portions of our CVS repository. We normally fork off a branch
- for a stable release just before starting the development cycle for
- the next release.
-
- The first thing you have to know is the branch name for the branch you
- are interested in getting at. To do this, look at some long-lived
- file, say the top-level HISTORY file, with "cvs status -v" to see what
- the branch names are. (Thanks to Ian Lance Taylor for pointing out
- that this is the easiest way to do it.) Typical branch names are:
- REL7_1_STABLE
- REL7_0_PATCHES
- REL6_5_PATCHES
-
- OK, so how do you do work on a branch? By far the best way is to
- create a separate checkout tree for the branch and do your work in
- that. Not only is that the easiest way to deal with CVS, but you
- really need to have the whole past tree available anyway to test your
- work. (And you *better* test your work. Never forget that dot-releases
- tend to go out with very little beta testing --- so whenever you
- commit an update to a stable branch, you'd better be doubly sure that
- it's correct.)
-
- Normally, to checkout the head branch, you just cd to the place you
- want to contain the toplevel "pgsql" directory and say
- cvs ... checkout pgsql
-
- To get a past branch, you cd to wherever you want it and say
- cvs ... checkout -r BRANCHNAME pgsql
-
- For example, just a couple days ago I did
- mkdir ~postgres/REL7_1
- cd ~postgres/REL7_1
- cvs ... checkout -r REL7_1_STABLE pgsql
-
- and now I have a maintenance copy of 7.1.*.
-
- When you've done a checkout in this way, the branch name is "sticky":
- CVS automatically knows that this directory tree is for the branch,
- and whenever you do "cvs update" or "cvs commit" in this tree, you'll
- fetch or store the latest version in the branch, not the head version.
- Easy as can be.
-
- So, if you have a patch that needs to apply to both the head and a
- recent stable branch, you have to make the edits and do the commit
- twice, once in your development tree and once in your stable branch
- tree. This is kind of a pain, which is why we don't normally fork the
- tree right away after a major release --- we wait for a dot-release or
- two, so that we won't have to double-patch the first wave of fixes.
-
- 1.17) Where can I get a copy of the SQL standards?
-
- There are three versions of the SQL standard: SQL-92, SQL:1999, and
- SQL:2003. They are endorsed by ANSI and ISO. Draft versions can be
- downloaded from:
- * SQL-92 http://www.contrib.andrew.cmu.edu/~shadow/sql/sql1992.txt
- * SQL:1999
- http://www.cse.iitb.ac.in/dbms/Data/Papers-Other/SQL1999/ansi-iso-
- 9075-2-1999.pdf
- * SQL:2003 http://www.wiscorp.com/sql_2003_standard.zip
-
- Some SQL standards web pages are:
- * http://troels.arvin.dk/db/rdbms/links/#standards
- * http://www.wiscorp.com/SQLStandards.html
- * http://www.contrib.andrew.cmu.edu/~shadow/sql.html#syntax (SQL-92)
- * http://dbs.uni-leipzig.de/en/lokal/standards.pdf (paper)
-
- 1.18) Where can I get technical assistance?
-
- Many technical questions held by those new to the code have been
- answered on the pgsql-hackers mailing list - the archives of which can
- be found at http://archives.postgresql.org/pgsql-hackers/.
-
- If you cannot find discussion or your particular question, feel free
- to put it to the list.
-
- Major contributors also answer technical questions, including
- questions about development of new features, on IRC at
- irc.freenode.net in the #postgresql channel.
-
- 1.19) How do I get involved in PostgreSQL web site development?
-
- PostgreSQL website development is discussed on the
- pgsql-www@postgresql.org mailing list. The is a project page where the
- source code is available at
- http://gborg.postgresql.org/project/pgweb/projdisplay.php , the code
- for the next version of the website is under the "portal" module.
-
- 1.20) Why haven't you replaced CVS with SVN, Git, Monotone, VSS, <insert your
- favorite SCMS here>?
-
- Currently the core developers see no SCMS that will provide enough
- benefit to outwiegh the pain involved in moving to a new SCMS. Typical
- problems that must be addressed by any new SCMS include:
- * Run natively on all of our supported platforms.
- * Integrate into the Buildfarm.
- * Import our entire CVS Repository while preserving complete
- history.
- * Allow for anonymous checkouts.
-
- Currently there is no intention for switching to a new SCMS until at
- least the end of the 8.4 development cycle sometime in late 2008. For
- more information please refer to the mailing list archives.
-
-Technical Questions
-
- 2.1) How do I efficiently access information in tables from the backend code?
-
- You first need to find the tuples(rows) you are interested in. There
- are two ways. First, SearchSysCache() and related functions allow you
- to query the system catalogs. This is the preferred way to access
- system tables, because the first call to the cache loads the needed
- rows, and future requests can return the results without accessing the
- base table. The caches use system table indexes to look up tuples. A
- list of available caches is located in
- src/backend/utils/cache/syscache.c.
- src/backend/utils/cache/lsyscache.c contains many column-specific
- cache lookup functions.
-
- The rows returned are cache-owned versions of the heap rows.
- Therefore, you must not modify or delete the tuple returned by
- SearchSysCache(). What you should do is release it with
- ReleaseSysCache() when you are done using it; this informs the cache
- that it can discard that tuple if necessary. If you neglect to call
- ReleaseSysCache(), then the cache entry will remain locked in the
- cache until end of transaction, which is tolerable but not very
- desirable.
-
- If you can't use the system cache, you will need to retrieve the data
- directly from the heap table, using the buffer cache that is shared by
- all backends. The backend automatically takes care of loading the rows
- into the buffer cache.
-
- Open the table with heap_open(). You can then start a table scan with
- heap_beginscan(), then use heap_getnext() and continue as long as
- HeapTupleIsValid() returns true. Then do a heap_endscan(). Keys can be
- assigned to the scan. No indexes are used, so all rows are going to be
- compared to the keys, and only the valid rows returned.
-
- You can also use heap_fetch() to fetch rows by block number/offset.
- While scans automatically lock/unlock rows from the buffer cache, with
- heap_fetch(), you must pass a Buffer pointer, and ReleaseBuffer() it
- when completed.
-
- Once you have the row, you can get data that is common to all tuples,
- like t_self and t_oid, by merely accessing the HeapTuple structure
- entries. If you need a table-specific column, you should take the
- HeapTuple pointer, and use the GETSTRUCT() macro to access the
- table-specific start of the tuple. You then cast the pointer as a
- Form_pg_proc pointer if you are accessing the pg_proc table, or
- Form_pg_type if you are accessing pg_type. You can then access the
- columns by using a structure pointer:
-((Form_pg_class) GETSTRUCT(tuple))->relnatts
-
- You must not directly change live tuples in this way. The best way is
- to use heap_modifytuple() and pass it your original tuple, and the
- values you want changed. It returns a palloc'ed tuple, which you pass
- to heap_replace(). You can delete tuples by passing the tuple's t_self
- to heap_destroy(). You use t_self for heap_update() too. Remember,
- tuples can be either system cache copies, which might go away after
- you call ReleaseSysCache(), or read directly from disk buffers, which
- go away when you heap_getnext(), heap_endscan, or ReleaseBuffer(), in
- the heap_fetch() case. Or it may be a palloc'ed tuple, that you must
- pfree() when finished.
-
- 2.2) Why are table, column, type, function, view names sometimes referenced
- as Name or NameData, and sometimes as char *?
-
- Table, column, type, function, and view names are stored in system
- tables in columns of type Name. Name is a fixed-length,
- null-terminated type of NAMEDATALEN bytes. (The default value for
- NAMEDATALEN is 64 bytes.)
-typedef struct nameData
- {
- char data[NAMEDATALEN];
- } NameData;
- typedef NameData *Name;
-
- Table, column, type, function, and view names that come into the
- backend via user queries are stored as variable-length,
- null-terminated character strings.
-
- Many functions are called with both types of names, ie. heap_open().
- Because the Name type is null-terminated, it is safe to pass it to a
- function expecting a char *. Because there are many cases where
- on-disk names(Name) are compared to user-supplied names(char *), there
- are many cases where Name and char * are used interchangeably.
-
- 2.3) Why do we use Node and List to make data structures?
-
- We do this because this allows a consistent way to pass data inside
- the backend in a flexible way. Every node has a NodeTag which
- specifies what type of data is inside the Node. Lists are groups of
- Nodes chained together as a forward-linked list.
-
- Here are some of the List manipulation commands:
-
- lfirst(i), lfirst_int(i), lfirst_oid(i)
- return the data (a pointer, integer or OID respectively) of
- list cell i.
-
- lnext(i)
- return the next list cell after i.
-
- foreach(i, list)
- loop through list, assigning each list cell to i. It is
- important to note that i is a ListCell *, not the data in the
- List element. You need to use lfirst(i) to get at the data.
- Here is a typical code snippet that loops through a List
- containing Var *'s and processes each one:
-
-
- List *list;
- ListCell *i;
-
- foreach(i, list)
- {
- Var *var = lfirst(i);
-
- /* process var here */
- }
-
- lcons(node, list)
- add node to the front of list, or create a new list with node
- if list is NIL.
-
- lappend(list, node)
- add node to the end of list.
-
- list_concat(list1, list2)
- Concatenate list2 on to the end of list1.
-
- list_length(list)
- return the length of the list.
-
- list_nth(list, i)
- return the i'th element in list, counting from zero.
-
- lcons_int, ...
- There are integer versions of these: lcons_int, lappend_int,
- etc. Also versions for OID lists: lcons_oid, lappend_oid, etc.
-
- You can print nodes easily inside gdb. First, to disable output
- truncation when you use the gdb print command:
-(gdb) set print elements 0
-
- Instead of printing values in gdb format, you can use the next two
- commands to print out List, Node, and structure contents in a verbose
- format that is easier to understand. List's are unrolled into nodes,
- and nodes are printed in detail. The first prints in a short format,
- and the second in a long format:
-(gdb) call print(any_pointer)
- (gdb) call pprint(any_pointer)
-
- The output appears in the postmaster log file, or on your screen if
- you are running a backend directly without a postmaster.
-
- 2.4) I just added a field to a structure. What else should I do?
-
- The structures passed around in the parser, rewriter, optimizer, and
- executor require quite a bit of support. Most structures have support
- routines in src/backend/nodes used to create, copy, read, and output
- those structures (in particular, the files copyfuncs.c and
- equalfuncs.c. Make sure you add support for your new field to these
- files. Find any other places the structure might need code for your
- new field. mkid is helpful with this (see 1.10).
-
- 2.5) Why do we use palloc() and pfree() to allocate memory?
-
- palloc() and pfree() are used in place of malloc() and free() because
- we find it easier to automatically free all memory allocated when a
- query completes. This assures us that all memory that was allocated
- gets freed even if we have lost track of where we allocated it. There
- are special non-query contexts that memory can be allocated in. These
- affect when the allocated memory is freed by the backend.
-
- 2.6) What is ereport()?
-
- ereport() is used to send messages to the front-end, and optionally
- terminate the current query being processed. The first parameter is an
- ereport level of DEBUG (levels 1-5), LOG, INFO, NOTICE, ERROR, FATAL,
- or PANIC. NOTICE prints on the user's terminal and the postmaster
- logs. INFO prints only to the user's terminal and LOG prints only to
- the server logs. (These can be changed from postgresql.conf.) ERROR
- prints in both places, and terminates the current query, never
- returning from the call. FATAL terminates the backend process. The
- remaining parameters of ereport are a printf-style set of parameters
- to print.
-
- ereport(ERROR) frees most memory and open file descriptors so you
- don't need to clean these up before the call.
-
- 2.7) What is CommandCounterIncrement()?
-
- Normally, transactions can not see the rows they modify. This allows
- UPDATE foo SET x = x + 1 to work correctly.
-
- However, there are cases where a transactions needs to see rows
- affected in previous parts of the transaction. This is accomplished
- using a Command Counter. Incrementing the counter allows transactions
- to be broken into pieces so each piece can see rows modified by
- previous pieces. CommandCounterIncrement() increments the Command
- Counter, creating a new part of the transaction.
-
- 2.8) What debugging features are available?
-
- First, try running configure with the --enable-cassert option, many
- assert()s monitor the progress of the backend and halt the program
- when something unexpected occurs.
-
- The postmaster has a -d option that allows even more detailed
- information to be reported. The -d option takes a number that
- specifies the debug level. Be warned that high debug level values
- generate large log files.
-
- If the postmaster is not running, you can actually run the postgres
- backend from the command line, and type your SQL statement directly.
- This is recommended only for debugging purposes. If you have compiled
- with debugging symbols, you can use a debugger to see what is
- happening. Because the backend was not started from postmaster, it is
- not running in an identical environment and locking/backend
- interaction problems might not be duplicated.
-
- If the postmaster is running, start psql in one window, then find the
- PID of the postgres process used by psql using SELECT
- pg_backend_pid(). Use a debugger to attach to the postgres PID. You
- can set breakpoints in the debugger and issue queries from the other.
- If you are looking to find the location that is generating an error or
- log message, set a breakpoint at errfinish. psql. If you are debugging
- postgres startup, you can set PGOPTIONS="-W n", then start psql. This
- will cause startup to delay for n seconds so you can attach to the
- process with the debugger, set any breakpoints, and continue through
- the startup sequence.
-
- You can also compile with profiling to see what functions are taking
- execution time. The backend profile files will be deposited in the
- pgsql/data directory. The client profile file will be put in the
- client's current directory. Linux requires a compile with
- -DLINUX_PROFILE for proper profiling.
+ http://wiki.postgresql.org/wiki/Development_information