Tom Lane [Sun, 30 May 1999 15:17:58 +0000 (15:17 +0000)]
Replace static rcsid[] strings by IDENTIFICATION comments in
file headers, to conform to established Postgres coding style and avoid
warnings from gcc.
Tom Lane [Sun, 30 May 1999 00:06:44 +0000 (00:06 +0000)]
configure.in forgot to do AC_LANG_C to reselect C-based tests
after checking for presence of C++ compiler. Odd we hadn't seen any
reports of problems before...
Vadim B. Mikheev [Sat, 29 May 1999 10:25:33 +0000 (10:25 +0000)]
1. Run all pg_dump queries in single serializable transaction.
2. Get rid of locking when updating statistics in vacuum.
3. Use QuerySnapshot in COPY TO and call SetQuerySnashot
in main tcop loop before FETCH and COPY TO.
Tom Lane [Sat, 29 May 1999 01:48:06 +0000 (01:48 +0000)]
new_relation_targetlist used to cause about 8 separate (and
redundant) SearchSysCache searches per table column in an INSERT, which
accounted for a good percentage of the CPU time for INSERT ... VALUES().
Now it only does two searches in the typical case.
Tom Lane [Fri, 28 May 1999 17:03:31 +0000 (17:03 +0000)]
Repair performance problem in SI segment manipulations: iterating
through MAXBACKENDS array entries used to be fine when MAXBACKENDS = 64.
It's not so cool with MAXBACKENDS = 1024 (or more!), especially not in a
frequently-used routine like SIDelExpiredDataEntries. Repair by making
procState array size be the soft MaxBackends limit rather than the hard
limit, and by converting SIGetProcStateLimit() to a macro.
Tom Lane [Fri, 28 May 1999 01:54:53 +0000 (01:54 +0000)]
When closure of the backend connection is detected during pqFlush,
do the right thing: look for a NOTICE message from the backend before we
close our side of the socket. 6.4 libpq did not reliably print the backend's
hara-kiri message, 'The Postmaster has informed me ...', because it only
did the right thing if connection closure was detected during a read
attempt instead of a write attempt.
Tom Lane [Wed, 26 May 1999 22:57:39 +0000 (22:57 +0000)]
Patch from Andreas: when CREATE TABLE is followed by CREATE INDEX
before any tuples are loaded, preserve the default '1000 tuples' table
size estimate.
Tom Lane [Wed, 26 May 1999 21:51:13 +0000 (21:51 +0000)]
Fix pg_dump to use the same maximum-query-size constant as
the backend does. Remove unnecessary limitation on field size in
dumpClasses_dumpData (ie, -d or -D case).
Add chapters on CVS access, MVCC, SQL theory to the docs.
Add an appendix with more details on date/time attributes and handling.
Update most references to Postgres version numbers to 6.5,
*except* for the porting list which will require a report
from a successful installation to be updated.
Vadim B. Mikheev [Tue, 25 May 1999 18:31:28 +0000 (18:31 +0000)]
Get rid of page-level locking in btree-s.
BT_READ/BT_WRITE are BUFFER_LOCK_SHARE/BUFFER_LOCK_EXCLUSIVE now.
Also get rid of #define BT_VERSION_1 - we use version 1 as default
for near two years now.
Vadim B. Mikheev [Tue, 25 May 1999 18:20:31 +0000 (18:20 +0000)]
Get rid of page-level locking in btree-s.
LockBuffer is used to acquire read/write access
to index pages. Pages are released before leaving
index internals.
Jan Wieck [Tue, 25 May 1999 13:16:10 +0000 (13:16 +0000)]
Bugfix - Range table entries that are unused after rewriting should
not be marked inFromCl any longer. Otherwise the planner gets confused
and joins over them where in fact it does not have to.
Adjust hasSubLinks now with a recursive lookup - could be wrong in
multi action rules because parse state isn't reset correctly and all
actions in the rule are marked hasSubLinks if one of them has.
Tom Lane [Sun, 23 May 1999 21:41:14 +0000 (21:41 +0000)]
Detect case of invalid use of GROUP BY when there are no
aggregate functions, as in
select a, b from foo group by a;
The ungrouped reference to b is not kosher, but formerly we neglected to
check this unless there was an aggregate function somewhere in the query.
Bruce Momjian [Sun, 23 May 1999 01:04:07 +0000 (01:04 +0000)]
Here it is. Remove or rename the current interfaces/libpq++ and untar
this file in interfaces/
It will all need to be checked in. I used the char *rcsid[] method for
cvs ids so it can be strings | grep'd to find version numbers. The new
version for the library is 3.0.
Run configure from src/ to create the Makefile and it should be good to
go.
I did minimal documentation references in the README, I'll see if I can
get something to Tom Lockhart rather quickly.
Tom Lane [Sat, 22 May 1999 23:19:37 +0000 (23:19 +0000)]
Modify aset.c logic so that blocks requested from malloc get
bigger the more space is used in an allocset. This reduces the malloc
overhead very substantially on queries that need lots of memory.
Tom Lane [Sat, 22 May 1999 17:47:54 +0000 (17:47 +0000)]
Modify backend switch parsing to prevent 'insecure' switches
from being accepted when they are passed from client connection request.
Get rid of a couple that no longer do anything (like -P).
Jan Wieck [Thu, 20 May 1999 12:12:55 +0000 (12:12 +0000)]
Fixed shift/reduce conflict
SelectStmt and CursorStmt tried to parse FOR UPDATE ... / FOR READ ONLY.
Cursor now checks that it is read only by looking at forUpdate of Query.
SelectStmt handles FOR READ ONLY too.
Rearrange and consolidate the Admin Guide.
Add reference pages for utilities and remove standalone chapters for same.
Add material for an appendix on date/time properties, but not yet
integrated with the User's Guide.
Break up the former chapter on pg_options
into Admin and Programmer's Guides.
Tom Lane [Thu, 20 May 1999 02:57:15 +0000 (02:57 +0000)]
Update CREATE FUNCTION docs: mention use of AS clause with
INTERNAL functions, add a warning about trying to overload function names
for dynamically loaded C functions (from old man page).
Tom Lane [Tue, 18 May 1999 23:40:05 +0000 (23:40 +0000)]
Add Aggref and ArrayRef to the set of node types that transformExpr
will pass through rather than spitting up. This is necessary to handle
cases where coerce_type causes a subexpression to be retransformed, as in
SELECT count(*) + 1.0 FROM table
Tom Lane [Tue, 18 May 1999 21:36:10 +0000 (21:36 +0000)]
Now that hashjoin is reliable for large joins (knock on wood),
remove optimizer's arbitrary limit on how large a join it will use hashing
for. (The limit was too large to prevent the problems we'd been seeing,
anyway...)