Re: Bad date format., posted by Stefan Ritt on Thu Dec 15 17:33:39 2005
|
Alex H wrote: |
[Liste]
Type Derničre image = Date
[Image Routeurs]
Type Derniere Image = Date
|
Maybe your additional accent. The attribute list contains Derniere Image, but in your first logbook you say Type Derničre image with accent grave, so this are not the same words. |
Re: Bad date format., posted by Alex H on Mon Dec 19 09:41:08 2005
|
Stefan Ritt wrote: |
Alex H wrote: |
[Liste]
Type Derničre image = Date
[Image Routeurs]
Type Derniere Image = Date
|
Maybe your additional accent. The attribute list contains Derniere Image, but in your first logbook you say Type Derničre image with accent grave, so this are not the same words. |
It's work now !
Thanks ! |
elog hand on destination mirror server, posted by stephane on Thu Feb 21 12:50:53 2008
|
Hello,
When I start elogd on the mirror destination server, until cron job started, I can access to the website normaly. Once cron job started, I can't access to the website of the elog mirror destination server.
How can I solve this problem ?
Kind regards
Stéphane |
Odd behaviour with gecko based browsers and reverse proxy., posted by Andraž 'ruskie' Levstik on Thu Mar 12 08:03:07 2009
|
I have a really odd issue with elog...
My setup:
elogd.cfg:
[global]
port = 2109
SMTP host = localhost
Resource dir = /usr/share/elog
Logbook dir = /srv/www/elog
charset = UTF-8
Main Tab = main
URL = https://elog.codemages.net
Relative redirection = 1
Usr = www
Grp = www
Resolve host names = 0
Self register = 0
Admin user = ruskie
SSL = 0
Login expiration = 20
[links]
Subdir = links
Theme = default
Comment = Link spam
Attributes = Author, Category, Link, Comment
Options Category = General, Law, IT, Odds, Linux
Required Attributes = Link, Comment
Page Title = ELOG - Link spam
Reverse sort = 1
Quick filter = Date, Category
Time format = %Y-%m-%dT%H:%M
Date format = %Y-%m-%d
Password file = links.pwd
Admin user = ruskie
RSS Title = $logbook $entry time :$Author :$Link :$Comment
[notes]
Subdir = notes
Theme = default
Comment = Various notes
Attributes = Category, Comment
Options Category = General, Law, IT, Odds, Linux
Required Attributes = Comment
Page Title = ELOG - Various notes
Reverse sort = 1
Quick filter = Date, Category
Time format = %Y-%m-%dT%H:%M
Date format = %Y-%m-%d
Password file = notes.pwd
Admin user = ruskie
RSS Title = $logbook $entry time :$Comment
lighttpd.conf:
$HTTP["host"] =~ "elog.*" {
proxy.server = ( "" =>
( "10.0.0.10" =>
(
"host" => "10.0.0.10",
"port" => 2109
)
)
)
}
If I access elog directly on 2109 I get no issues but when I use this configuration it takes minutes to access
any elog page at all. But the odd thing is this only happens when I'm using a gecko based browser(firefox,
xulrunner based etc...) if I use webkit-gtk or elinks all the pages load up with no noticable delay.
I have tried this on different computers, different configurations and so on... And always the same behaviour
when it comes to a gecko based browser. Elinks and webkit-gtk load up fine.
Marking this as Other so far. If anyone knows anything about running elog and lighttpd togheter please speak up. |
Move to: elog crashes with large no of entries being moved., posted by David Pilgram on Wed Jun 10 13:56:09 2009
|
Hi Stefan,
I've been slowly moving threads, and twice now so far (and reproducably) had elog crash.
In each case, it is trying to move a thread with more than 24 entries; it copies the first 24 entries, then
crashes with "Segmentation Fault". It does not erase the lock file /var/run/elog.pid
I have got around this by manually copying the entries beyond no 24, then deleting the thread entry by entry.
I am aware that I have an old and limited machine (586, 256MB RAM, running Slack 10), and at first I was
"content" to write it off as that; but when it crashed for the second time at exactly the same entry (the
twenty-forth) even though the size of the entries would have been significantly different, I wondered if there
was some factor within elog that could affect this.
I've not tried it with Copy to:, but imagine it will also be affected as the only difference with this and Move
to: is the deletion of the thread after all the entries had been copied. |
Re: Move to: elog crashes with large no of entries being moved., posted by Stefan Ritt on Wed Jun 10 14:09:04 2009
|
> Hi Stefan,
>
> I've been slowly moving threads, and twice now so far (and reproducably) had elog crash.
>
> In each case, it is trying to move a thread with more than 24 entries; it copies the first 24 entries, then
> crashes with "Segmentation Fault". It does not erase the lock file /var/run/elog.pid
>
> I have got around this by manually copying the entries beyond no 24, then deleting the thread entry by entry.
>
> I am aware that I have an old and limited machine (586, 256MB RAM, running Slack 10), and at first I was
> "content" to write it off as that; but when it crashed for the second time at exactly the same entry (the
> twenty-forth) even though the size of the entries would have been significantly different, I wondered if there
> was some factor within elog that could affect this.
>
> I've not tried it with Copy to:, but imagine it will also be affected as the only difference with this and Move
> to: is the deletion of the thread after all the entries had been copied.
This rings a bell: it's probably related to some internal stack overflow, since the entries are copied
recursively. I have an idea on how to fix that, but I need time for that. |
Re: Move to: elog crashes with large no of entries being moved., posted by David Pilgram on Wed Jun 10 15:31:13 2009
|
> > Hi Stefan,
> >
> > I've been slowly moving threads, and twice now so far (and reproducably) had elog crash.
> >
> > In each case, it is trying to move a thread with more than 24 entries; it copies the first 24 entries, then
> > crashes with "Segmentation Fault". It does not erase the lock file /var/run/elog.pid
> >
> > I have got around this by manually copying the entries beyond no 24, then deleting the thread entry by entry.
> >
> > I am aware that I have an old and limited machine (586, 256MB RAM, running Slack 10), and at first I was
> > "content" to write it off as that; but when it crashed for the second time at exactly the same entry (the
> > twenty-forth) even though the size of the entries would have been significantly different, I wondered if there
> > was some factor within elog that could affect this.
> >
> > I've not tried it with Copy to:, but imagine it will also be affected as the only difference with this and Move
> > to: is the deletion of the thread after all the entries had been copied.
>
> This rings a bell: it's probably related to some internal stack overflow, since the entries are copied
> recursively. I have an idea on how to fix that, but I need time for that.
Thanks Stefan, I'll be keeping an eye out on any annoucement about this one! |
Re: Move to: elog crashes with large no of entries being moved., posted by Stefan Ritt on Thu Jun 25 15:55:04 2009
|
> Hi Stefan,
>
> I've been slowly moving threads, and twice now so far (and reproducably) had elog crash.
>
> In each case, it is trying to move a thread with more than 24 entries; it copies the first 24 entries, then
> crashes with "Segmentation Fault". It does not erase the lock file /var/run/elog.pid
>
> I have got around this by manually copying the entries beyond no 24, then deleting the thread entry by entry.
>
> I am aware that I have an old and limited machine (586, 256MB RAM, running Slack 10), and at first I was
> "content" to write it off as that; but when it crashed for the second time at exactly the same entry (the
> twenty-forth) even though the size of the entries would have been significantly different, I wondered if there
> was some factor within elog that could affect this.
>
> I've not tried it with Copy to:, but imagine it will also be affected as the only difference with this and Move
> to: is the deletion of the thread after all the entries had been copied.
I reworked the internal memory allocation, since there was a stack overflow going over 24 entries. It should be now
much better. Give a try to revision 2226. |
|