tech-kern archive
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index][Old Index]
Re: CVS commit: src/sys/uvm
On Wed, Dec 22, 2010 at 05:37:57AM +0000, YAMAMOTO Takashi wrote:
> hi,
>
> > Could you ack this discussion?
>
> sorry for dropping a ball.
>
> >
> > On Tue, Dec 07, 2010 at 01:19:46AM +0900, Masao Uebayashi wrote:
> >> On Thu, Nov 25, 2010 at 11:32:39PM +0000, YAMAMOTO Takashi wrote:
> >> > [ adding cc: tech-kern@ ]
> >> >
> >> > hi,
> >> >
> >> > > On Wed, Nov 24, 2010 at 11:26:39PM -0800, Matt Thomas wrote:
> >> > >>
> >> > >> On Nov 24, 2010, at 10:47 PM, Masao Uebayashi wrote:
> >> > >>
> >> > >> > On Thu, Nov 25, 2010 at 05:44:21AM +0000, YAMAMOTO Takashi wrote:
> >> > >> >> hi,
> >> > >> >>
> >> > >> >>> On Thu, Nov 25, 2010 at 04:18:25AM +0000, YAMAMOTO Takashi wrote:
> >> > >> >>>> hi,
> >> > >> >>>>
> >> > >> >>>>> Hi, thanks for review.
> >> > >> >>>>>
> >> > >> >>>>> On Thu, Nov 25, 2010 at 01:58:04AM +0000, YAMAMOTO Takashi
> >> > >> >>>>> wrote:
> >> > >> >>>>>> hi,
> >> > >> >>>>>>
> >> > >> >>>>>> - what's VM_PHYSSEG_OP_PG?
> >> > >> >>>>>
> >> > >> >>>>> It's to lookup vm_physseg by "struct vm_page *", relying on that
> >> > >> >>>>> "struct vm_page *[]" is allocated linearly. It'll be used to
> >> > >> >>>>> remove
> >> > >> >>>>> vm_page::phys_addr as we talked some time ago.
> >> > >> >>>>
> >> > >> >>>> i'm not sure if commiting this unused uncommented code now helps
> >> > >> >>>> it.
> >> > >> >>>> some try-and-benchmark cycles might be necessary given that
> >> > >> >>>> vm_page <-> paddr conversion could be performace critical.
> >> > >> >>>
> >> > >> >>> If you really care performance, we can directly pass "struct
> >> > >> >>> vm_page
> >> > >> >>> *" to pmap_enter().
> >> > >> >>>
> >> > >> >>> We're doing "struct vm_page *" -> "paddr_t" just before
> >> > >> >>> pmap_enter(),
> >> > >> >>> then doing "paddr_t" -> "vm_physseg" reverse lookup again in
> >> > >> >>> pmap_enter() to check if a given PA is managed. What is really
> >> > >> >>> needed here is, to lookup "struct vm_page *" -> "vm_physseg" once
> >> > >> >>> and you'll know both paddr_t and managed or not.
> >> > >> >>
> >> > >> >> i agree that the current code is not ideal in that respect.
> >> > >> >> otoh, i'm not sure if passing vm_physseg around is a good idea.
> >> > >> >
> >> > >> > It's great you share the interest.
> >> > >> >
> >> > >> > I chose vm_physseg, because it was there. I'm open to alternatives,
> >> > >> > but I don't think you have many options...
> >> > >>
> >> > >> Passing vm_page * doesn't work if the page isn't managed since there
> >> > >> won't be a vm_page for the paddr_t.
> >> > >>
> >> > >> Now passing both paddr_t and vm_page * would work and if the pointer
> >> > >> to the vm_page, it would be an unmanaged mapping. This also gives the
> >> > >> access to mdpg without another lookup.
> >> > >
> >> > > What if XIP'ed md(4), where physical pages are in .data (or .rodata)?
> >> > >
> >> > > And don't forget that you're the one who first pointed out that
> >> > > allocating vm_pages for XIP is a pure waste of memory. ;)
> >> >
> >> > i guess matt meant "if the pointer to the vm_page is NULL,".
> >> >
> >> > >
> >> > > I'm allocating vm_pages, only because of phys_addr and loan_count.
> >> > > I believe vm_pages is unnecessary for read-only XIP segments.
> >> > > Because they're read-only, and stateless.
> >> > >
> >> > > I've already concluded that the current "managed or not" model
> >> > > doesn't work for XIP. I'm pretty sure that my vm_physseg + off_t
> >> > > model can explain everything. I'm rather waiting for a counter
> >> > > example how vm_physseg doesn't work...
> >> >
> >> > i guess your suggestion is too vague.
> >> > where do you want to use vm_physseg * + off_t instead of vm_page * ?
> >> > getpages, pmap_enter, and? how their function prototypes would be?
> >>
> >> The basic idea is straightforward; always allocate vm_physseg for
> >> memories/devices. If a vm_physseg is used as general purpose
> >> memory, you allocate vm_page[] (as vm_physseg::pgs). If it's
> >> potentially mapped as cached, you allocate pvh (as vm_physseg:pvh).
>
> can you provide function prototypes?
I have no real code for this big picture at this moment. Making
vm_physseg available as reference is the first step. This only
changes uvm_page_physload() to return a pointer:
-void uvm_page_physload();
+void *uvm_page_physload();
But this makes XIP pager MUCH cleaner. The reason has been explained
many times.
Making fault handlers and pagers to use vm_physseg * + off_t is
the next step, and I don't intend to work on it now. I just want
to explain the big picture.
>
> >>
> >> Keep vm_physseg * + off_t array on stack. If UVM objects uses
> >> vm_page (e.g. vnode), its pager looks up vm_page -> vm_physseg *
> >> + off_t *once* and cache it on stack.
>
> do you mean something like this?
> struct {
> vm_physseg *hoge;
> off_t fuga;
> } foo [16];
Yes.
Or cache vm_page * with it, like:
struct vm_id {
vm_physseg *seg;
off_t off;
vm_page *pg;
};
uvm_fault()
{
vm_id pgs[];
:
}
Vnode pager (genfs_getpages) takes vm_page's by looking up
vnode::v_uobj's list, or uvn_findpages().
When it returns back to fault handler, we have to lookup vm_physseg
for each page. Then fill the "seg" slot above (assuming we'll
remove vm_page::phys_addr soon).
Fault handler calls per-vm_page operations iff vm_page slot is filled.
XIP pages are not pageq'ed. XIP pages don't need vm_page, but
cached because it's vnode.
(Just in case, have you read my paper?)
Home |
Main Index |
Thread Index |
Old Index